F S UID          PID    PPID     LWP  C NLWP PRI  NI ADDR SZ WCHAN  STIME TTY          TIME CMD
4 S root           1       0       1  1    1  80   0 - 47301 ep_pol 06:53 ?        00:02:29 /usr/lib/systemd/systemd --switched-root --system --deserialize 28
1 S root           2       0       2  0    1  80   0 -     0 kthrea 06:53 ?        00:00:00 [kthreadd]
1 I root           3       2       3  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [rcu_gp]
1 I root           4       2       4  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [rcu_par_gp]
1 I root           5       2       5  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [slub_flushwq]
1 I root           6       2       6  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [netns]
1 I root           8       2       8  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/0:0H-events_highpri]
1 I root          10       2      10  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [mm_percpu_wq]
1 I root          12       2      12  0    1  80   0 -     0 rcu_ta 06:53 ?        00:00:00 [rcu_tasks_kthre]
1 I root          13       2      13  0    1  80   0 -     0 rcu_ta 06:53 ?        00:00:00 [rcu_tasks_rude_]
1 I root          14       2      14  0    1  80   0 -     0 rcu_ta 06:53 ?        00:00:00 [rcu_tasks_trace]
1 S root          15       2      15  0    1  80   0 -     0 smpboo 06:53 ?        00:00:02 [ksoftirqd/0]
1 I root          16       2      16  0    1  80   0 -     0 rcu_gp 06:53 ?        00:00:25 [rcu_preempt]
1 S root          17       2      17  0    1 -40   - -     0 smpboo 06:53 ?        00:00:00 [migration/0]
1 S root          18       2      18  0    1   9   - -     0 smpboo 06:53 ?        00:00:00 [idle_inject/0]
1 S root          20       2      20  0    1  80   0 -     0 smpboo 06:53 ?        00:00:00 [cpuhp/0]
1 S root          21       2      21  0    1  80   0 -     0 smpboo 06:53 ?        00:00:00 [cpuhp/1]
1 S root          22       2      22  0    1   9   - -     0 smpboo 06:53 ?        00:00:00 [idle_inject/1]
1 S root          23       2      23  0    1 -40   - -     0 smpboo 06:53 ?        00:00:00 [migration/1]
1 S root          24       2      24  0    1  80   0 -     0 smpboo 06:53 ?        00:00:01 [ksoftirqd/1]
1 I root          26       2      26  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/1:0H-events_highpri]
1 S root          27       2      27  0    1  80   0 -     0 smpboo 06:53 ?        00:00:00 [cpuhp/2]
1 S root          28       2      28  0    1   9   - -     0 smpboo 06:53 ?        00:00:00 [idle_inject/2]
1 S root          29       2      29  0    1 -40   - -     0 smpboo 06:53 ?        00:00:00 [migration/2]
1 S root          30       2      30  0    1  80   0 -     0 smpboo 06:53 ?        00:00:01 [ksoftirqd/2]
1 I root          32       2      32  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/2:0H-events_highpri]
1 S root          33       2      33  0    1  80   0 -     0 smpboo 06:53 ?        00:00:00 [cpuhp/3]
1 S root          34       2      34  0    1   9   - -     0 smpboo 06:53 ?        00:00:00 [idle_inject/3]
1 S root          35       2      35  0    1 -40   - -     0 smpboo 06:53 ?        00:00:00 [migration/3]
1 S root          36       2      36  0    1  80   0 -     0 smpboo 06:53 ?        00:00:01 [ksoftirqd/3]
1 I root          38       2      38  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/3:0H-events_highpri]
1 S root          39       2      39  0    1  80   0 -     0 smpboo 06:53 ?        00:00:00 [cpuhp/4]
1 S root          40       2      40  0    1   9   - -     0 smpboo 06:53 ?        00:00:00 [idle_inject/4]
1 S root          41       2      41  0    1 -40   - -     0 smpboo 06:53 ?        00:00:00 [migration/4]
1 S root          42       2      42  0    1  80   0 -     0 smpboo 06:53 ?        00:00:01 [ksoftirqd/4]
1 I root          44       2      44  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/4:0H-events_highpri]
1 S root          45       2      45  0    1  80   0 -     0 smpboo 06:53 ?        00:00:00 [cpuhp/5]
1 S root          46       2      46  0    1   9   - -     0 smpboo 06:53 ?        00:00:00 [idle_inject/5]
1 S root          47       2      47  0    1 -40   - -     0 smpboo 06:53 ?        00:00:00 [migration/5]
1 S root          48       2      48  0    1  80   0 -     0 smpboo 06:53 ?        00:00:05 [ksoftirqd/5]
1 I root          50       2      50  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/5:0H-events_highpri]
1 S root          51       2      51  0    1  80   0 -     0 smpboo 06:53 ?        00:00:00 [cpuhp/6]
1 S root          52       2      52  0    1   9   - -     0 smpboo 06:53 ?        00:00:00 [idle_inject/6]
1 S root          53       2      53  0    1 -40   - -     0 smpboo 06:53 ?        00:00:00 [migration/6]
1 S root          54       2      54  0    1  80   0 -     0 smpboo 06:53 ?        00:00:01 [ksoftirqd/6]
1 I root          56       2      56  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/6:0H-events_highpri]
1 S root          57       2      57  0    1  80   0 -     0 smpboo 06:53 ?        00:00:00 [cpuhp/7]
1 S root          58       2      58  0    1   9   - -     0 smpboo 06:53 ?        00:00:00 [idle_inject/7]
1 S root          59       2      59  0    1 -40   - -     0 smpboo 06:53 ?        00:00:00 [migration/7]
1 S root          60       2      60  0    1  80   0 -     0 smpboo 06:53 ?        00:00:01 [ksoftirqd/7]
1 I root          62       2      62  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/7:0H-events_highpri]
1 S root          63       2      63  0    1  80   0 -     0 smpboo 06:53 ?        00:00:00 [cpuhp/8]
1 S root          64       2      64  0    1   9   - -     0 smpboo 06:53 ?        00:00:00 [idle_inject/8]
1 S root          65       2      65  0    1 -40   - -     0 smpboo 06:53 ?        00:00:00 [migration/8]
1 S root          66       2      66  0    1  80   0 -     0 smpboo 06:53 ?        00:00:01 [ksoftirqd/8]
1 I root          68       2      68  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/8:0H-events_highpri]
1 S root          69       2      69  0    1  80   0 -     0 smpboo 06:53 ?        00:00:00 [cpuhp/9]
1 S root          70       2      70  0    1   9   - -     0 smpboo 06:53 ?        00:00:00 [idle_inject/9]
1 S root          71       2      71  0    1 -40   - -     0 smpboo 06:53 ?        00:00:00 [migration/9]
1 S root          72       2      72  0    1  80   0 -     0 smpboo 06:53 ?        00:00:01 [ksoftirqd/9]
1 I root          74       2      74  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/9:0H-events_highpri]
1 S root          75       2      75  0    1  80   0 -     0 smpboo 06:53 ?        00:00:00 [cpuhp/10]
1 S root          76       2      76  0    1   9   - -     0 smpboo 06:53 ?        00:00:00 [idle_inject/10]
1 S root          77       2      77  0    1 -40   - -     0 smpboo 06:53 ?        00:00:00 [migration/10]
1 S root          78       2      78  0    1  80   0 -     0 smpboo 06:53 ?        00:00:01 [ksoftirqd/10]
1 I root          80       2      80  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/10:0H-events_highpri]
1 S root          81       2      81  0    1  80   0 -     0 smpboo 06:53 ?        00:00:00 [cpuhp/11]
1 S root          82       2      82  0    1   9   - -     0 smpboo 06:53 ?        00:00:00 [idle_inject/11]
1 S root          83       2      83  0    1 -40   - -     0 smpboo 06:53 ?        00:00:00 [migration/11]
1 S root          84       2      84  0    1  80   0 -     0 smpboo 06:53 ?        00:00:01 [ksoftirqd/11]
1 I root          86       2      86  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/11:0H-events_highpri]
5 S root          98       2      98  0    1  80   0 -     0 devtmp 06:53 ?        00:00:00 [kdevtmpfs]
1 I root          99       2      99  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [inet_frag_wq]
1 S root         100       2     100  0    1  80   0 -     0 kaudit 06:53 ?        00:00:00 [kauditd]
1 S root         101       2     101  0    1  80   0 -     0 watchd 06:53 ?        00:00:00 [khungtaskd]
1 S root         102       2     102  0    1  80   0 -     0 oom_re 06:53 ?        00:00:00 [oom_reaper]
1 I root         103       2     103  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [writeback]
1 S root         104       2     104  0    1  80   0 -     0 kcompa 06:53 ?        00:00:05 [kcompactd0]
1 S root         105       2     105  0    1  85   5 -     0 ksm_sc 06:53 ?        00:00:00 [ksmd]
1 S root         106       2     106  0    1  99  19 -     0 khugep 06:53 ?        00:00:01 [khugepaged]
1 I root         107       2     107  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [cryptd]
1 I root         108       2     108  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [kintegrityd]
1 I root         109       2     109  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [kblockd]
1 I root         110       2     110  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [blkcg_punt_bio]
1 I root         111       2     111  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [tpm_dev_wq]
1 I root         112       2     112  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [md]
1 I root         113       2     113  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [md_bitmap]
1 I root         114       2     114  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [edac-poller]
1 S root         115       2     115  0    1   9   - -     0 kthrea 06:53 ?        00:00:00 [watchdogd]
1 I root         116       2     116  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/0:1H-xfs-log/vda4]
1 S root         118       2     118  0    1  80   0 -     0 kswapd 06:53 ?        00:00:12 [kswapd0]
1 I root         125       2     125  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [kthrotld]
1 I root         134       2     134  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [acpi_thermal_pm]
1 S root         135       2     135  0    1  80   0 -     0 hwrng_ 06:53 ?        00:00:00 [hwrng]
1 I root         136       2     136  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [kmpath_rdacd]
1 I root         137       2     137  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [kaluad]
1 I root         139       2     139  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [mld]
1 I root         140       2     140  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/7:1H-xfs-log/vda4]
1 I root         141       2     141  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [ipv6_addrconf]
1 I root         150       2     150  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [kstrp]
1 I root         160       2     160  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [zswap-shrink]
1 I root         161       2     161  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/u25:0]
1 I root         272       2     272  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/1:1H-xfs-log/vda4]
1 I root         392       2     392  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/2:1H-xfs-log/vda4]
1 I root         394       2     394  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/6:1H-xfs-log/vda4]
1 I root         396       2     396  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/9:1H-kblockd]
1 I root         398       2     398  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/10:1H-xfs-log/vda4]
1 I root         400       2     400  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/8:1H-kblockd]
1 I root         403       2     403  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/3:1H-xfs-log/vda4]
1 I root         404       2     404  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/11:1H-xfs-log/vda4]
1 I root         408       2     408  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/4:1H-xfs-log/vda4]
1 I root         412       2     412  0    1  60 -20 -     0 worker 06:53 ?        00:00:00 [kworker/5:1H-kblockd]
1 I root         505       2     505  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [iscsi_conn_clea]
1 I root         508       2     508  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [tls-strp]
1 I root         509       2     509  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [cnic_wq]
1 S root         510       2     510  0    1  60 -20 -     0 bnx2i_ 06:53 ?        00:00:00 [bnx2i_thread/0]
1 S root         511       2     511  0    1  60 -20 -     0 bnx2i_ 06:53 ?        00:00:00 [bnx2i_thread/1]
1 S root         512       2     512  0    1  60 -20 -     0 bnx2i_ 06:53 ?        00:00:00 [bnx2i_thread/2]
1 S root         513       2     513  0    1  60 -20 -     0 bnx2i_ 06:53 ?        00:00:00 [bnx2i_thread/3]
1 S root         514       2     514  0    1  60 -20 -     0 bnx2i_ 06:53 ?        00:00:00 [bnx2i_thread/4]
1 S root         515       2     515  0    1  60 -20 -     0 bnx2i_ 06:53 ?        00:00:00 [bnx2i_thread/5]
1 S root         516       2     516  0    1  60 -20 -     0 bnx2i_ 06:53 ?        00:00:00 [bnx2i_thread/6]
1 S root         517       2     517  0    1  60 -20 -     0 bnx2i_ 06:53 ?        00:00:00 [bnx2i_thread/7]
1 S root         518       2     518  0    1  60 -20 -     0 bnx2i_ 06:53 ?        00:00:00 [bnx2i_thread/8]
1 S root         519       2     519  0    1  60 -20 -     0 bnx2i_ 06:53 ?        00:00:00 [bnx2i_thread/9]
1 S root         520       2     520  0    1  60 -20 -     0 bnx2i_ 06:53 ?        00:00:00 [bnx2i_thread/10]
1 S root         521       2     521  0    1  60 -20 -     0 bnx2i_ 06:53 ?        00:00:00 [bnx2i_thread/11]
1 I root         712       2     712  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [ata_sff]
1 S root         714       2     714  0    1  80   0 -     0 scsi_e 06:53 ?        00:00:00 [scsi_eh_0]
1 I root         715       2     715  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [scsi_tmf_0]
1 S root         716       2     716  0    1  80   0 -     0 scsi_e 06:53 ?        00:00:00 [scsi_eh_1]
1 I root         717       2     717  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [scsi_tmf_1]
1 I root         718       2     718  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [nvme-wq]
1 I root         719       2     719  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [nvme-reset-wq]
1 I root         720       2     720  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [nvme-delete-wq]
1 I root         721       2     721  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [nvme-auth-wq]
1 I root         724       2     724  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [nvme_tcp_wq]
1 I root         764       2     764  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [kmpathd]
1 I root         765       2     765  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [kmpath_handlerd]
1 I root         779       2     779  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [xfsalloc]
1 I root         780       2     780  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [xfs_mru_cache]
1 I root         781       2     781  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [xfs-buf/vda4]
1 I root         782       2     782  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [xfs-conv/vda4]
1 I root         783       2     783  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [xfs-reclaim/vda]
1 I root         784       2     784  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [xfs-blockgc/vda]
1 I root         785       2     785  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [xfs-inodegc/vda]
1 I root         786       2     786  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [xfs-log/vda4]
1 I root         787       2     787  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [xfs-cil/vda4]
1 S root         788       2     788  0    1  80   0 -     0 xfsail 06:53 ?        00:00:05 [xfsaild/vda4]
4 S root         886       1     886  0    1  80   0 - 17022 ep_pol 06:53 ?        00:00:45 /usr/lib/systemd/systemd-journald
4 S root         897       1     897  0    1  80   0 -  8888 ep_pol 06:53 ?        00:00:06 /usr/lib/systemd/systemd-udevd
1 S root         998       2     998  0    1  80   0 -     0 kjourn 06:53 ?        00:00:00 [jbd2/vda3-8]
1 I root         999       2     999  0    1  60 -20 -     0 rescue 06:53 ?        00:00:00 [ext4-rsv-conver]
5 S root        1008       1    1008  0    2  76  -4 -  5396 do_sel 06:55 ?        00:00:00 /sbin/auditd
1 S root        1008       1    1009  0    2  76  -4 -  5396 futex_ 06:55 ?        00:00:00 /sbin/auditd
4 S dbus        1032       1    1032  0    1  80   0 -  4382 ep_pol 06:55 ?        00:00:00 /usr/bin/dbus-broker-launch --scope system --audit
4 S dbus        1034    1032    1034  0    1  80   0 -  1366 ep_pol 06:55 ?        00:00:43 dbus-broker --log 4 --controller 9 --machine-id 21801e6708c44f15b81395eb736a7cec --max-bytes 536870912 --max-fds 4096 --max-matches 131072 --audit
4 S root        1048       1    1048  0    2  80   0 - 19811 do_pol 06:55 ?        00:00:00 /usr/sbin/irqbalance --foreground
1 S root        1048       1    1055  0    2  80   0 - 19811 do_pol 06:55 ?        00:00:00 /usr/sbin/irqbalance --foreground
4 S root        1066       1    1066  0    1  80   0 -  5260 ep_pol 06:55 ?        00:00:02 /usr/lib/systemd/systemd-logind
5 S chrony      1073       1    1073  0    1  80   0 - 22710 do_sel 06:55 ?        00:00:00 /usr/sbin/chronyd -F 2
1 S openvsw+    1143       1    1143  0    1  70 -10 -  6252 do_pol 06:55 ?        00:00:39 ovsdb-server /etc/openvswitch/conf.db -vconsole:emer -vsyslog:err -vfile:info --remote=punix:/var/run/openvswitch/db.sock --private-key=db:Open_vSwitch,SSL,private_key --certificate=db:Open_vSwitch,SSL,certificate --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovsdb-server.log --pidfile=/var/run/openvswitch/ovsdb-server.pid --detach
1 I root        1151       2    1151  0    1  60 -20 -     0 rescue 06:55 ?        00:00:00 [ib-comp-wq]
1 I root        1152       2    1152  0    1  60 -20 -     0 rescue 06:55 ?        00:00:00 [ib-comp-unb-wq]
1 I root        1153       2    1153  0    1  60 -20 -     0 rescue 06:55 ?        00:00:00 [ib_mcast]
1 I root        1154       2    1154  0    1  60 -20 -     0 rescue 06:55 ?        00:00:00 [ib_nl_sa_wq]
5 S openvsw+    1210       1    1210  1   18  70 -10 - 352426 do_pol 06:55 ?       00:01:55 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    1222  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:11 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4329  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:03 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4330  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:02 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4331  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:04 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4332  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:03 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4333  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:03 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4334  0   18  70 -10 - 352426 do_pol 06:55 ?       00:01:11 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4335  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:02 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4336  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:03 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4337  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:03 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4338  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:03 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4339  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:02 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4340  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:02 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4341  1   18  70 -10 - 352426 do_pol 06:55 ?       00:01:58 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4342  0   18  70 -10 - 352426 do_pol 06:55 ?       00:00:55 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4343  0   18  70 -10 - 352426 do_pol 06:55 ?       00:01:06 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
5 S openvsw+    1210       1    4344  0   18  70 -10 - 352426 do_pol 06:55 ?       00:01:15 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
4 S root        1248       1    1248  1    3  80   0 - 65794 do_pol 06:55 ?        00:01:29 /usr/sbin/NetworkManager --no-daemon
1 S root        1248       1    1249  0    3  80   0 - 65794 do_pol 06:55 ?        00:00:00 /usr/sbin/NetworkManager --no-daemon
1 S root        1248       1    1250  0    3  80   0 - 65794 do_pol 06:55 ?        00:00:35 /usr/sbin/NetworkManager --no-daemon
1 S root        1298       1    1298  0    2  80   0 - 19633 do_pol 06:55 ?        00:00:00 /usr/bin/rhsmcertd
1 S root        1298       1    1299  0    2  80   0 - 19633 do_pol 06:55 ?        00:00:00 /usr/bin/rhsmcertd
4 S root        1300       1    1300  0    1  80   0 -  4491 do_sel 06:55 ?        00:00:00 sshd: /usr/sbin/sshd -D [listener] 1 of 10-100 startups
4 S root        1304       1    1304  0    1  80   0 -   777 do_sel 06:55 tty1     00:00:34 /sbin/agetty -o -p -- \u --noclear - linux
4 S core        2078       1    2078  1    1  80   0 -  8522 ep_pol 06:55 ?        00:01:57 /usr/lib/systemd/systemd --user
5 S core        2080    2078    2080  0    1  80   0 -  6918 do_sig 06:55 ?        00:00:00 (sd-pam)
1 S core        2156       1    2156  0    4  80   0 - 228404 inet_c 06:55 ?       00:00:00 /usr/bin/python3.9
1 S core        2156       1  261527  0    4  80   0 - 228404 do_sel 09:13 ?       00:00:00 /usr/bin/python3.9
1 S core        2156       1  261918  0    4  80   0 - 228404 do_sel 09:13 ?       00:00:00 /usr/bin/python3.9
1 S core        2156       1  262237  0    4  80   0 - 228404 do_sel 09:14 ?       00:00:00 /usr/bin/python3.9
4 S rpc         3520       1    3520  0    1  80   0 -  3317 do_pol 06:55 ?        00:00:00 /usr/bin/rpcbind -w -f
5 S rpcuser     3530       1    3530  0    1  80   0 -  2437 do_sel 06:55 ?        00:00:00 /usr/sbin/rpc.statd
4 S root        4764       1    4764  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:12 /usr/bin/crio
1 S root        4764       1    4765  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:58 /usr/bin/crio
5 S root        4764       1    4766  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:12 /usr/bin/crio
5 S root        4764       1    4767  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:01 /usr/bin/crio
5 S root        4764       1    4768  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:00 /usr/bin/crio
1 S root        4764       1    4769  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:00 /usr/bin/crio
5 S root        4764       1    4770  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:16 /usr/bin/crio
5 S root        4764       1    4771  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:15 /usr/bin/crio
5 S root        4764       1    4772  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:01 /usr/bin/crio
5 S root        4764       1    4773  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:33 /usr/bin/crio
5 S root        4764       1    4774  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:12 /usr/bin/crio
5 S root        4764       1    4775  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:21 /usr/bin/crio
5 S root        4764       1    4777  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:00 /usr/bin/crio
5 S root        4764       1    4794  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:01 /usr/bin/crio
5 S root        4764       1    4797  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:01 /usr/bin/crio
5 S root        4764       1    4838  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:01 /usr/bin/crio
5 S root        4764       1    4852  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:20 /usr/bin/crio
5 S root        4764       1    4876  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:01 /usr/bin/crio
5 S root        4764       1    4877  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:01 /usr/bin/crio
5 S root        4764       1    4878  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:00 /usr/bin/crio
5 S root        4764       1    4879  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:17 /usr/bin/crio
5 S root        4764       1    4880  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:01 /usr/bin/crio
5 S root        4764       1    4881  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:12 /usr/bin/crio
5 S root        4764       1    4882  0   52  80   0 - 1299962 futex_ 06:56 ?      00:00:21 /usr/bin/crio
5 S root        4764       1    8051  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:20 /usr/bin/crio
5 S root        4764       1    8052  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:25 /usr/bin/crio
5 S root        4764       1    8053  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:00 /usr/bin/crio
5 S root        4764       1    8054  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:32 /usr/bin/crio
5 S root        4764       1    8141  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:00 /usr/bin/crio
5 S root        4764       1    8181  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:16 /usr/bin/crio
5 S root        4764       1    8236  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:00 /usr/bin/crio
5 S root        4764       1    8239  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:30 /usr/bin/crio
5 S root        4764       1    8240  0   52  80   0 - 1299962 do_wai 06:58 ?      00:00:21 /usr/bin/crio
5 S root        4764       1    8241  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:00 /usr/bin/crio
5 S root        4764       1    8243  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:23 /usr/bin/crio
5 S root        4764       1    8273  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:00 /usr/bin/crio
5 S root        4764       1    8278  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:05 /usr/bin/crio
5 S root        4764       1    8349  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:15 /usr/bin/crio
5 S root        4764       1    8414  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:19 /usr/bin/crio
5 S root        4764       1    8650  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:21 /usr/bin/crio
5 S root        4764       1    8651  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:17 /usr/bin/crio
5 S root        4764       1    8652  0   52  80   0 - 1299962 do_wai 06:58 ?      00:00:20 /usr/bin/crio
5 S root        4764       1    8653  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:00 /usr/bin/crio
5 S root        4764       1    8664  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:25 /usr/bin/crio
5 S root        4764       1    8971  0   52  80   0 - 1299962 do_wai 06:58 ?      00:00:23 /usr/bin/crio
5 S root        4764       1    9016  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:00 /usr/bin/crio
5 S root        4764       1    9017  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:00 /usr/bin/crio
5 S root        4764       1    9018  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:10 /usr/bin/crio
5 S root        4764       1    9019  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:16 /usr/bin/crio
5 S root        4764       1    9020  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:13 /usr/bin/crio
5 S root        4764       1    9067  0   52  80   0 - 1299962 futex_ 06:58 ?      00:00:16 /usr/bin/crio
5 S root        4764       1    9068  0   52  80   0 - 1299962 ep_pol 06:58 ?      00:00:17 /usr/bin/crio
4 S root        4810       1    4810  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
1 S root        4810       1    4811  1   60  80   0 - 1482232 futex_ 06:56 ?      00:01:26 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
1 S root        4810       1    4812  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4813  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:02 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4814  0   60  80   0 - 1482232 wait_w 06:56 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
1 S root        4810       1    4815  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4816  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:13 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4817  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:03 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4818  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:13 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4819  0   60  80   0 - 1482232 futex_ 06:56 ?      00:01:22 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4820  1   60  80   0 - 1482232 futex_ 06:56 ?      00:01:39 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4821  0   60  80   0 - 1482232 futex_ 06:56 ?      00:01:22 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4822  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4823  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:16 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4825  1   60  80   0 - 1482232 futex_ 06:56 ?      00:01:27 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4826  0   60  80   0 - 1482232 futex_ 06:56 ?      00:01:16 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4827  0   60  80   0 - 1482232 ep_pol 06:56 ?      00:00:02 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4828  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:38 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4833  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:58 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    4853  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:18 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5376  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5377  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:14 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5378  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:06 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5379  1   60  80   0 - 1482232 futex_ 06:56 ?      00:01:37 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5380  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:15 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
1 S root        4810       1    5381  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5382  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5383  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:05 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5384  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5385  0   60  80   0 - 1482232 futex_ 06:56 ?      00:01:07 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5386  1   60  80   0 - 1482232 futex_ 06:56 ?      00:01:23 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5387  0   60  80   0 - 1482232 futex_ 06:56 ?      00:01:16 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
1 S root        4810       1    5388  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
1 S root        4810       1    5389  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5390  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:10 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5391  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:55 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5392  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:13 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5393  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:44 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1    5394  0   60  80   0 - 1482232 futex_ 06:56 ?      00:00:17 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70590  0   60  80   0 - 1482232 wait_w 07:20 ?      00:00:55 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70591  1   60  80   0 - 1482232 futex_ 07:20 ?      00:01:31 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
1 S root        4810       1   70592  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70602  1   60  80   0 - 1482232 futex_ 07:20 ?      00:01:18 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70603  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70604  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70605  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:16 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70606  1   60  80   0 - 1482232 futex_ 07:20 ?      00:01:18 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70608  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:02 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70620  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:49 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70621  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70623  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70626  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:02 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70631  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:01 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70632  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70633  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
5 S root        4810       1   70634  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:48 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
1 S root        4810       1   70635  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
1 S root        4810       1   70636  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
1 S root        4810       1   70637  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
1 S root        4810       1   70638  0   60  80   0 - 1482232 futex_ 07:20 ?      00:00:00 /usr/bin/kubelet --config=/etc/kubernetes/kubelet.conf --bootstrap-kubeconfig=/etc/kubernetes/kubeconfig --kubeconfig=/var/lib/kubelet/kubeconfig --container-runtime-endpoint=/var/run/crio/crio.sock --runtime-cgroups=/system.slice/crio.service --node-labels=node-role.kubernetes.io/control-plane,node-role.kubernetes.io/master,node.openshift.io/os_id=rhcos --node-ip=192.168.126.11 --minimum-container-ttl-duration=6m0s --cloud-provider= --volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --hostname-override= --provider-id= --register-with-taints=node-role.kubernetes.io/master=:NoSchedule --pod-infra-container-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d --system-reserved=cpu=200m,memory=350Mi,ephemeral-storage=350Mi --v=2
1 S root        4888       1    4888  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/0cdb480f634d8093290f0027899bf463b3d4da4cf9d2426709ec13b1c8879e0d/userdata -c 0cdb480f634d8093290f0027899bf463b3d4da4cf9d2426709ec13b1c8879e0d --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log --log-level info -n k8s_kube-controller-manager_kube-controller-manager-crc_openshift-kube-controller-manager_f614b9022728cf315e60c057852e563e_0 -P /run/containers/storage/overlay-containers/0cdb480f634d8093290f0027899bf463b3d4da4cf9d2426709ec13b1c8879e0d/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/0cdb480f634d8093290f0027899bf463b3d4da4cf9d2426709ec13b1c8879e0d/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/0cdb480f634d8093290f0027899bf463b3d4da4cf9d2426709ec13b1c8879e0d/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 0cdb480f634d8093290f0027899bf463b3d4da4cf9d2426709ec13b1c8879e0d -s
4 S root        4894    4888    4894  0   18  80   0 - 696079 ep_pol 06:56 ?       00:00:17 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4942  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:52 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4943  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:15 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4944  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:18 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4945  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:18 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4946  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:00 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4956  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:12 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4957  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:00 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4958  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:09 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4959  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:13 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4969  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:14 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4970  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:14 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4978  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:16 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4979  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:11 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    4989  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:00 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    5352  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:13 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    5353  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:16 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4894    4888    5354  0   18  80   0 - 696079 futex_ 06:56 ?       00:00:10 kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.217.0.0/22 --cluster-name=crc-rzkf7 --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=8760h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=false --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=10.217.4.0/23 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        4947       1    4947  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/a560786b579c49d9ce46fb91ef45ec643aae3891690c4c868ea855ca3f3e1fa9/userdata -c a560786b579c49d9ce46fb91ef45ec643aae3891690c4c868ea855ca3f3e1fa9 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log --log-level info -n k8s_cluster-policy-controller_kube-controller-manager-crc_openshift-kube-controller-manager_f614b9022728cf315e60c057852e563e_0 -P /run/containers/storage/overlay-containers/a560786b579c49d9ce46fb91ef45ec643aae3891690c4c868ea855ca3f3e1fa9/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/a560786b579c49d9ce46fb91ef45ec643aae3891690c4c868ea855ca3f3e1fa9/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/a560786b579c49d9ce46fb91ef45ec643aae3891690c4c868ea855ca3f3e1fa9/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u a560786b579c49d9ce46fb91ef45ec643aae3891690c4c868ea855ca3f3e1fa9 -s
4 S root        4949    4947    4949  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:00 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4960  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:31 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4961  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:01 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4962  0   18  80   0 - 673781 ep_pol 06:56 ?       00:00:03 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
5 S root        4949    4947    4963  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:12 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4964  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:00 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4974  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:10 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4975  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:01 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4976  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:00 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4977  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:10 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4980  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:00 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4981  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:02 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4982  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:01 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4994  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:10 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    4995  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:12 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    5019  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:12 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947    5402  0   18  80   0 - 673781 futex_ 06:56 ?       00:00:10 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4949    4947   28932  0   18  80   0 - 673781 futex_ 07:09 ?       00:00:00 cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2
1 S root        4966       1    4966  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/3070738247cb2a272c43f7a07fc9d92425ccf10b152e9a07f28099738d9cf46e/userdata -c 3070738247cb2a272c43f7a07fc9d92425ccf10b152e9a07f28099738d9cf46e --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager-cert-syncer/0.log --log-level info -n k8s_kube-controller-manager-cert-syncer_kube-controller-manager-crc_openshift-kube-controller-manager_f614b9022728cf315e60c057852e563e_0 -P /run/containers/storage/overlay-containers/3070738247cb2a272c43f7a07fc9d92425ccf10b152e9a07f28099738d9cf46e/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/3070738247cb2a272c43f7a07fc9d92425ccf10b152e9a07f28099738d9cf46e/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/3070738247cb2a272c43f7a07fc9d92425ccf10b152e9a07f28099738d9cf46e/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 3070738247cb2a272c43f7a07fc9d92425ccf10b152e9a07f28099738d9cf46e -s
4 S root        4968    4966    4968  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966    4983  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:01 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966    4984  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
5 S root        4968    4966    4985  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966    4986  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966    4987  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966    4988  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966    5011  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966    5012  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
5 S root        4968    4966    5013  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966    5350  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966    5351  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966    6322  0   16  80   0 - 603878 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966    6323  0   16  80   0 - 603878 ep_pol 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966   22624  0   16  80   0 - 603878 futex_ 07:06 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4968    4966   75512  0   16  80   0 - 603878 futex_ 07:22 ?       00:00:00 cluster-kube-controller-manager-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        4991       1    4991  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/cef89a639820f9c84213350e10d3c9d60db4072438772b16a98219a946110774/userdata -c cef89a639820f9c84213350e10d3c9d60db4072438772b16a98219a946110774 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager-recovery-controller/0.log --log-level info -n k8s_kube-controller-manager-recovery-controller_kube-controller-manager-crc_openshift-kube-controller-manager_f614b9022728cf315e60c057852e563e_0 -P /run/containers/storage/overlay-containers/cef89a639820f9c84213350e10d3c9d60db4072438772b16a98219a946110774/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/cef89a639820f9c84213350e10d3c9d60db4072438772b16a98219a946110774/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/cef89a639820f9c84213350e10d3c9d60db4072438772b16a98219a946110774/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u cef89a639820f9c84213350e10d3c9d60db4072438772b16a98219a946110774 -s
4 S root        4993    4991    4993  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    5002  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:02 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    5003  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
5 S root        4993    4991    5004  0   17  80   0 - 622439 ep_pol 06:56 ?       00:00:01 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    5005  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:01 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    5006  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    5007  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    5008  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    5009  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    5010  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
5 S root        4993    4991    5014  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:01 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    6435  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    6436  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    6437  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    6438  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    6439  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:01 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        4993    4991    6440  0   17  80   0 - 622439 futex_ 06:56 ?       00:00:00 cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2
1 S root        5016       1    5016  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/c6ee80464125edb57275df5080d958b19a4423fab16e5e62e4be550c84a514a4/userdata -c c6ee80464125edb57275df5080d958b19a4423fab16e5e62e4be550c84a514a4 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-machine-config-operator_kube-rbac-proxy-crio-crc_d1b160f5dda77d281dd8e69ec8d817f9/kube-rbac-proxy-crio/0.log --log-level info -n k8s_kube-rbac-proxy-crio_kube-rbac-proxy-crio-crc_openshift-machine-config-operator_d1b160f5dda77d281dd8e69ec8d817f9_0 -P /run/containers/storage/overlay-containers/c6ee80464125edb57275df5080d958b19a4423fab16e5e62e4be550c84a514a4/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/c6ee80464125edb57275df5080d958b19a4423fab16e5e62e4be550c84a514a4/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/c6ee80464125edb57275df5080d958b19a4423fab16e5e62e4be550c84a514a4/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u c6ee80464125edb57275df5080d958b19a4423fab16e5e62e4be550c84a514a4 -s
4 S nfsnobo+    5018    5016    5018  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
1 S nfsnobo+    5018    5016    5049  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
5 S nfsnobo+    5018    5016    5050  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
1 S nfsnobo+    5018    5016    5051  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
5 S nfsnobo+    5018    5016    5052  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
5 S nfsnobo+    5018    5016    5053  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
1 S nfsnobo+    5018    5016    5054  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
5 S nfsnobo+    5018    5016    5071  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
5 S nfsnobo+    5018    5016    5072  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
1 S nfsnobo+    5018    5016    5073  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
5 S nfsnobo+    5018    5016    5088  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
1 S nfsnobo+    5018    5016    5089  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
1 S nfsnobo+    5018    5016    5090  0   14  80   0 - 559288 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
5 S nfsnobo+    5018    5016    5100  0   14  80   0 - 559288 ep_pol 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=:9637 --config-file=/etc/kubernetes/crio-metrics-proxy.cfg --client-ca-file=/etc/kubernetes/kubelet-ca.crt --logtostderr=true --kubeconfig=/var/lib/kubelet/kubeconfig --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:9537 --tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem --tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem
1 S root        5021       1    5021  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/6e08332b0cf6ac85c2589d3bec3fae5de8fae8dea7f8790241ef5bfcbfcc797f/userdata -c 6e08332b0cf6ac85c2589d3bec3fae5de8fae8dea7f8790241ef5bfcbfcc797f --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-scheduler_openshift-kube-scheduler-crc_3dcd261975c3d6b9a6ad6367fd4facd3/kube-scheduler/0.log --log-level info -n k8s_kube-scheduler_openshift-kube-scheduler-crc_openshift-kube-scheduler_3dcd261975c3d6b9a6ad6367fd4facd3_0 -P /run/containers/storage/overlay-containers/6e08332b0cf6ac85c2589d3bec3fae5de8fae8dea7f8790241ef5bfcbfcc797f/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/6e08332b0cf6ac85c2589d3bec3fae5de8fae8dea7f8790241ef5bfcbfcc797f/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/6e08332b0cf6ac85c2589d3bec3fae5de8fae8dea7f8790241ef5bfcbfcc797f/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 6e08332b0cf6ac85c2589d3bec3fae5de8fae8dea7f8790241ef5bfcbfcc797f -s
4 S root        5024    5021    5024  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:02 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5060  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:09 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5061  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:03 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5062  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:03 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5063  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:01 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5064  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:00 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5083  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:00 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5084  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:00 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5097  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:00 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5098  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:04 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5099  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:03 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5130  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:00 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5131  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:00 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5167  0   17  80   0 - 619303 ep_pol 06:56 ?       00:00:01 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5168  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:03 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5169  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:00 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5024    5021    5198  0   17  80   0 - 619303 futex_ 06:56 ?       00:00:02 kube-scheduler --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --cert-dir=/var/run/kubernetes --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=false,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false -v=2 --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S root        5030       1    5030  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/420e661db0423df4e0c9a5879239e1867cdc6bc279f0ccae9fc82f4a76842d57/userdata -c 420e661db0423df4e0c9a5879239e1867cdc6bc279f0ccae9fc82f4a76842d57 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver/0.log --log-level info -n k8s_kube-apiserver_kube-apiserver-crc_openshift-kube-apiserver_f4b27818a5e8e43d0dc095d08835c792_0 -P /run/containers/storage/overlay-containers/420e661db0423df4e0c9a5879239e1867cdc6bc279f0ccae9fc82f4a76842d57/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/420e661db0423df4e0c9a5879239e1867cdc6bc279f0ccae9fc82f4a76842d57/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/420e661db0423df4e0c9a5879239e1867cdc6bc279f0ccae9fc82f4a76842d57/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 420e661db0423df4e0c9a5879239e1867cdc6bc279f0ccae9fc82f4a76842d57 -s
4 S root        5036    5030    5036  0   11  80   0 - 501820 futex_ 06:56 ?       00:00:00 watch-termination --termination-touch-file=/var/log/kube-apiserver/.terminating --termination-log-file=/var/log/kube-apiserver/termination.log --graceful-termination-duration=15s --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig -- hyperkube kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5036    5030    5055  0   11  80   0 - 501820 futex_ 06:56 ?       00:00:00 watch-termination --termination-touch-file=/var/log/kube-apiserver/.terminating --termination-log-file=/var/log/kube-apiserver/termination.log --graceful-termination-duration=15s --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig -- hyperkube kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5036    5030    5056  0   11  80   0 - 501820 do_wai 06:56 ?       00:00:00 watch-termination --termination-touch-file=/var/log/kube-apiserver/.terminating --termination-log-file=/var/log/kube-apiserver/termination.log --graceful-termination-duration=15s --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig -- hyperkube kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5036    5030    5057  0   11  80   0 - 501820 ep_pol 06:56 ?       00:00:00 watch-termination --termination-touch-file=/var/log/kube-apiserver/.terminating --termination-log-file=/var/log/kube-apiserver/termination.log --graceful-termination-duration=15s --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig -- hyperkube kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5036    5030    5058  0   11  80   0 - 501820 futex_ 06:56 ?       00:00:00 watch-termination --termination-touch-file=/var/log/kube-apiserver/.terminating --termination-log-file=/var/log/kube-apiserver/termination.log --graceful-termination-duration=15s --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig -- hyperkube kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5036    5030    5059  0   11  80   0 - 501820 futex_ 06:56 ?       00:00:00 watch-termination --termination-touch-file=/var/log/kube-apiserver/.terminating --termination-log-file=/var/log/kube-apiserver/termination.log --graceful-termination-duration=15s --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig -- hyperkube kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5036    5030    5080  0   11  80   0 - 501820 futex_ 06:56 ?       00:00:00 watch-termination --termination-touch-file=/var/log/kube-apiserver/.terminating --termination-log-file=/var/log/kube-apiserver/termination.log --graceful-termination-duration=15s --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig -- hyperkube kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5036    5030    5081  0   11  80   0 - 501820 futex_ 06:56 ?       00:00:00 watch-termination --termination-touch-file=/var/log/kube-apiserver/.terminating --termination-log-file=/var/log/kube-apiserver/termination.log --graceful-termination-duration=15s --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig -- hyperkube kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5036    5030    5082  0   11  80   0 - 501820 futex_ 06:56 ?       00:00:00 watch-termination --termination-touch-file=/var/log/kube-apiserver/.terminating --termination-log-file=/var/log/kube-apiserver/termination.log --graceful-termination-duration=15s --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig -- hyperkube kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5036    5030    6535  0   11  80   0 - 501820 futex_ 06:56 ?       00:00:00 watch-termination --termination-touch-file=/var/log/kube-apiserver/.terminating --termination-log-file=/var/log/kube-apiserver/termination.log --graceful-termination-duration=15s --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig -- hyperkube kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5036    5030   11934  0   11  80   0 - 501820 futex_ 06:59 ?       00:00:00 watch-termination --termination-touch-file=/var/log/kube-apiserver/.terminating --termination-log-file=/var/log/kube-apiserver/termination.log --graceful-termination-duration=15s --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig -- hyperkube kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5066       1    5066  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/5a95445ed0d63749e461fc0e86d27b474e0e374fab9e603019b88c7efcdd7fe3/userdata -c 5a95445ed0d63749e461fc0e86d27b474e0e374fab9e603019b88c7efcdd7fe3 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-scheduler_openshift-kube-scheduler-crc_3dcd261975c3d6b9a6ad6367fd4facd3/kube-scheduler-cert-syncer/0.log --log-level info -n k8s_kube-scheduler-cert-syncer_openshift-kube-scheduler-crc_openshift-kube-scheduler_3dcd261975c3d6b9a6ad6367fd4facd3_0 -P /run/containers/storage/overlay-containers/5a95445ed0d63749e461fc0e86d27b474e0e374fab9e603019b88c7efcdd7fe3/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/5a95445ed0d63749e461fc0e86d27b474e0e374fab9e603019b88c7efcdd7fe3/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/5a95445ed0d63749e461fc0e86d27b474e0e374fab9e603019b88c7efcdd7fe3/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 5a95445ed0d63749e461fc0e86d27b474e0e374fab9e603019b88c7efcdd7fe3 -s
4 S root        5068    5066    5068  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5091  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:02 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5092  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5093  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5094  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5095  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5096  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5142  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5144  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5145  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5153  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5154  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
5 S root        5068    5066    5165  0   16  80   0 - 603511 ep_pol 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5894  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066    5895  0   16  80   0 - 603511 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5068    5066   74725  0   16  80   0 - 603511 futex_ 07:21 ?       00:00:00 cluster-kube-scheduler-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5076       1    5076  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/5228c086bc36fd866c1499f082f903419fa3673334236a07bff20bcfce43b06d/userdata -c 5228c086bc36fd866c1499f082f903419fa3673334236a07bff20bcfce43b06d --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log --log-level info -n k8s_kube-apiserver-cert-syncer_kube-apiserver-crc_openshift-kube-apiserver_f4b27818a5e8e43d0dc095d08835c792_0 -P /run/containers/storage/overlay-containers/5228c086bc36fd866c1499f082f903419fa3673334236a07bff20bcfce43b06d/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/5228c086bc36fd866c1499f082f903419fa3673334236a07bff20bcfce43b06d/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/5228c086bc36fd866c1499f082f903419fa3673334236a07bff20bcfce43b06d/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 5228c086bc36fd866c1499f082f903419fa3673334236a07bff20bcfce43b06d -s
4 S root        5078    5076    5078  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    5101  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    5102  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    5103  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    5105  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    5106  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    5149  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    5150  0   16  80   0 - 605033 ep_pol 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    5151  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    5158  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
5 S root        5078    5076    5159  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    5160  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    6408  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    6409  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076    6410  0   16  80   0 - 605033 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
1 S root        5078    5076  100085  0   16  80   0 - 605033 futex_ 07:46 ?       00:00:00 cluster-kube-apiserver-operator cert-syncer --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver --destination-dir=/etc/kubernetes/static-pod-certs
4 S root        5087    5036    5087  0   21  80   0 - 1483517 futex_ 06:56 ?      00:00:00 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5111  2   21  80   0 - 1483517 hrtime 06:56 ?      00:03:05 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5112  3   21  80   0 - 1483517 futex_ 06:56 ?      00:04:55 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
5 S root        5087    5036    5113  2   21  80   0 - 1483517 futex_ 06:56 ?      00:03:42 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5114  0   21  80   0 - 1483517 futex_ 06:56 ?      00:00:00 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5115  0   21  80   0 - 1483517 futex_ 06:56 ?      00:00:00 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5161  3   21  80   0 - 1483517 futex_ 06:56 ?      00:04:21 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5170  3   21  80   0 - 1483517 futex_ 06:56 ?      00:04:28 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5171  3   21  80   0 - 1483517 futex_ 06:56 ?      00:04:45 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5172  3   21  80   0 - 1483517 futex_ 06:56 ?      00:04:28 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5228  3   21  80   0 - 1483517 futex_ 06:56 ?      00:04:57 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5229  1   21  80   0 - 1483517 futex_ 06:56 ?      00:02:24 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
5 S root        5087    5036    5230  3   21  80   0 - 1483517 futex_ 06:56 ?      00:04:14 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5231  3   21  80   0 - 1483517 futex_ 06:56 ?      00:04:55 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5232  2   21  80   0 - 1483517 futex_ 06:56 ?      00:03:18 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5233  3   21  80   0 - 1483517 ep_pol 06:56 ?      00:04:35 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5234  3   21  80   0 - 1483517 futex_ 06:56 ?      00:04:37 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5349  3   21  80   0 - 1483517 futex_ 06:56 ?      00:05:06 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036    5770  3   21  80   0 - 1483517 futex_ 06:56 ?      00:04:55 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036   29503  0   21  80   0 - 1483517 futex_ 07:10 ?      00:00:51 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5087    5036   70470  1   21  80   0 - 1483517 futex_ 07:20 ?      00:02:00 kube-apiserver --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml --advertise-address=192.168.126.11 -v=2 --permit-address-sharing
1 S root        5107       1    5107  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/5e4747b628060a1275ffd6a56e6b8fbecf053fc56b96672837b0042116a03c4a/userdata -c 5e4747b628060a1275ffd6a56e6b8fbecf053fc56b96672837b0042116a03c4a --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-scheduler_openshift-kube-scheduler-crc_3dcd261975c3d6b9a6ad6367fd4facd3/kube-scheduler-recovery-controller/0.log --log-level info -n k8s_kube-scheduler-recovery-controller_openshift-kube-scheduler-crc_openshift-kube-scheduler_3dcd261975c3d6b9a6ad6367fd4facd3_0 -P /run/containers/storage/overlay-containers/5e4747b628060a1275ffd6a56e6b8fbecf053fc56b96672837b0042116a03c4a/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/5e4747b628060a1275ffd6a56e6b8fbecf053fc56b96672837b0042116a03c4a/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/5e4747b628060a1275ffd6a56e6b8fbecf053fc56b96672837b0042116a03c4a/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 5e4747b628060a1275ffd6a56e6b8fbecf053fc56b96672837b0042116a03c4a -s
4 S root        5109    5107    5109  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    5124  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:01 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    5125  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    5126  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    5127  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    5129  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    5140  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:01 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    5141  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    5143  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
5 S root        5109    5107    5152  0   17  80   0 - 621816 ep_pol 06:56 ?       00:00:01 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
5 S root        5109    5107    5180  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    5217  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    5218  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    5219  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    5220  0   17  80   0 - 621816 futex_ 06:56 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    6596  0   17  80   0 - 621816 futex_ 06:57 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5109    5107    6597  0   17  80   0 - 621816 futex_ 06:57 ?       00:00:00 cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-scheduler --listen=0.0.0.0:11443 -v=2
1 S root        5122       1    5122  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/3c2e4a800fbf060433677fe1ebedef1ea2bfe9ab42e9707ea0e1802a14c24abd/userdata -c 3c2e4a800fbf060433677fe1ebedef1ea2bfe9ab42e9707ea0e1802a14c24abd --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-regeneration-controller/0.log --log-level info -n k8s_kube-apiserver-cert-regeneration-controller_kube-apiserver-crc_openshift-kube-apiserver_f4b27818a5e8e43d0dc095d08835c792_0 -P /run/containers/storage/overlay-containers/3c2e4a800fbf060433677fe1ebedef1ea2bfe9ab42e9707ea0e1802a14c24abd/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/3c2e4a800fbf060433677fe1ebedef1ea2bfe9ab42e9707ea0e1802a14c24abd/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/3c2e4a800fbf060433677fe1ebedef1ea2bfe9ab42e9707ea0e1802a14c24abd/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 3c2e4a800fbf060433677fe1ebedef1ea2bfe9ab42e9707ea0e1802a14c24abd -s
4 S root        5128    5122    5128  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122    5135  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:02 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122    5136  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122    5137  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122    5138  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122    5139  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122    5146  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122    5147  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
5 S root        5128    5122    5148  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
5 S root        5128    5122    5155  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
5 S root        5128    5122    5156  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122    5224  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122    5225  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122    5226  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122    5364  0   17  80   0 - 623690 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122   10329  0   17  80   0 - 623690 ep_pol 06:58 ?       00:00:01 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5128    5122   48774  0   17  80   0 - 623690 futex_ 07:16 ?       00:00:00 cluster-kube-apiserver-operator cert-regeneration-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-apiserver-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-apiserver -v=2
1 S root        5162       1    5162  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/3a6eb1b4994bffe9eab51a1c97c6327858f5ce305eea9b78f472742f51f4abf6/userdata -c 3a6eb1b4994bffe9eab51a1c97c6327858f5ce305eea9b78f472742f51f4abf6 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-insecure-readyz/0.log --log-level info -n k8s_kube-apiserver-insecure-readyz_kube-apiserver-crc_openshift-kube-apiserver_f4b27818a5e8e43d0dc095d08835c792_0 -P /run/containers/storage/overlay-containers/3a6eb1b4994bffe9eab51a1c97c6327858f5ce305eea9b78f472742f51f4abf6/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/3a6eb1b4994bffe9eab51a1c97c6327858f5ce305eea9b78f472742f51f4abf6/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/3a6eb1b4994bffe9eab51a1c97c6327858f5ce305eea9b78f472742f51f4abf6/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 3a6eb1b4994bffe9eab51a1c97c6327858f5ce305eea9b78f472742f51f4abf6 -s
4 S root        5164    5162    5164  0   11  80   0 - 512612 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator insecure-readyz --insecure-port=6080 --delegate-url=https://localhost:6443/readyz
1 S root        5164    5162    5175  0   11  80   0 - 512612 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator insecure-readyz --insecure-port=6080 --delegate-url=https://localhost:6443/readyz
1 S root        5164    5162    5176  0   11  80   0 - 512612 ep_pol 06:56 ?       00:00:00 cluster-kube-apiserver-operator insecure-readyz --insecure-port=6080 --delegate-url=https://localhost:6443/readyz
1 S root        5164    5162    5177  0   11  80   0 - 512612 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator insecure-readyz --insecure-port=6080 --delegate-url=https://localhost:6443/readyz
1 S root        5164    5162    5178  0   11  80   0 - 512612 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator insecure-readyz --insecure-port=6080 --delegate-url=https://localhost:6443/readyz
1 S root        5164    5162    5179  0   11  80   0 - 512612 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator insecure-readyz --insecure-port=6080 --delegate-url=https://localhost:6443/readyz
1 S root        5164    5162    5181  0   11  80   0 - 512612 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator insecure-readyz --insecure-port=6080 --delegate-url=https://localhost:6443/readyz
1 S root        5164    5162    5182  0   11  80   0 - 512612 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator insecure-readyz --insecure-port=6080 --delegate-url=https://localhost:6443/readyz
5 S root        5164    5162    5183  0   11  80   0 - 512612 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator insecure-readyz --insecure-port=6080 --delegate-url=https://localhost:6443/readyz
1 S root        5164    5162    5184  0   11  80   0 - 512612 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator insecure-readyz --insecure-port=6080 --delegate-url=https://localhost:6443/readyz
1 S root        5164    5162    5185  0   11  80   0 - 512612 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator insecure-readyz --insecure-port=6080 --delegate-url=https://localhost:6443/readyz
1 S root        5236       1    5236  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/1e9c4fabba8c14b97af3c9e4c5a06530342ad9551c3c72ed70d09f0303cde1c4/userdata -c 1e9c4fabba8c14b97af3c9e4c5a06530342ad9551c3c72ed70d09f0303cde1c4 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-etcd_etcd-crc_2139d3e2895fc6797b9c76a1b4c9886d/etcdctl/0.log --log-level info -n k8s_etcdctl_etcd-crc_openshift-etcd_2139d3e2895fc6797b9c76a1b4c9886d_0 -P /run/containers/storage/overlay-containers/1e9c4fabba8c14b97af3c9e4c5a06530342ad9551c3c72ed70d09f0303cde1c4/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/1e9c4fabba8c14b97af3c9e4c5a06530342ad9551c3c72ed70d09f0303cde1c4/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/1e9c4fabba8c14b97af3c9e4c5a06530342ad9551c3c72ed70d09f0303cde1c4/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 1e9c4fabba8c14b97af3c9e4c5a06530342ad9551c3c72ed70d09f0303cde1c4 -s
4 S root        5238    5236    5238  0    1  80   0 -  1047 do_wai 06:56 ?        00:00:00 /bin/bash -c trap TERM INT; sleep infinity & wait
4 S root        5242    5238    5242  0    1  80   0 -  1233 do_sys 06:56 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep infinity
1 S root        5244       1    5244  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/0e5ef2fcb17d33a46afa178f70754b021c2962c9ec91e99830d622b07b102d31/userdata -c 0e5ef2fcb17d33a46afa178f70754b021c2962c9ec91e99830d622b07b102d31 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-etcd_etcd-crc_2139d3e2895fc6797b9c76a1b4c9886d/etcd/0.log --log-level info -n k8s_etcd_etcd-crc_openshift-etcd_2139d3e2895fc6797b9c76a1b4c9886d_0 -P /run/containers/storage/overlay-containers/0e5ef2fcb17d33a46afa178f70754b021c2962c9ec91e99830d622b07b102d31/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/0e5ef2fcb17d33a46afa178f70754b021c2962c9ec91e99830d622b07b102d31/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/0e5ef2fcb17d33a46afa178f70754b021c2962c9ec91e99830d622b07b102d31/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 0e5ef2fcb17d33a46afa178f70754b021c2962c9ec91e99830d622b07b102d31 -s
4 S root        5246    5244    5246  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:00:00 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5333  2   19  61 -19 - 3104770 futex_ 06:56 ?      00:03:08 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5334  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:00:08 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5335  0   19  61 -19 - 3104770 ep_pol 06:56 ?      00:01:03 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5336  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:01:08 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5337  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:00:00 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5338  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:00:58 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5339  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:01:10 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5340  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:00:00 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5341  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:01:07 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5342  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:01:04 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5343  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:01:00 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5344  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:01:04 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5345  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:01:00 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5346  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:00:51 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5347  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:00:55 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244    5348  0   19  61 -19 - 3104770 futex_ 06:56 ?      00:01:04 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244   10231  0   19  61 -19 - 3104770 futex_ 06:58 ?      00:01:01 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5246    5244   32989  0   19  61 -19 - 3104770 futex_ 07:12 ?      00:00:53 etcd --logger=zap --log-level=info --experimental-initial-corrupt-check=true --snapshot-count=10000 --initial-advertise-peer-urls=https://192.168.126.11:2380 --cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --client-cert-auth=true --peer-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --peer-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --peer-trusted-ca-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --peer-client-cert-auth=true --advertise-client-urls=https://192.168.126.11:2379 --listen-client-urls=https://0.0.0.0:2379,unixs://192.168.126.11:0 --listen-peer-urls=https://0.0.0.0:2380 --metrics=extensive --listen-metrics-urls=https://0.0.0.0:9978
1 S root        5261       1    5261  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/ccac1833655d5c2436c7564aa95846abaa866b1db78a1cdf3db84fff990e7e8a/userdata -c ccac1833655d5c2436c7564aa95846abaa866b1db78a1cdf3db84fff990e7e8a --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-etcd_etcd-crc_2139d3e2895fc6797b9c76a1b4c9886d/etcd-metrics/0.log --log-level info -n k8s_etcd-metrics_etcd-crc_openshift-etcd_2139d3e2895fc6797b9c76a1b4c9886d_0 -P /run/containers/storage/overlay-containers/ccac1833655d5c2436c7564aa95846abaa866b1db78a1cdf3db84fff990e7e8a/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/ccac1833655d5c2436c7564aa95846abaa866b1db78a1cdf3db84fff990e7e8a/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/ccac1833655d5c2436c7564aa95846abaa866b1db78a1cdf3db84fff990e7e8a/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u ccac1833655d5c2436c7564aa95846abaa866b1db78a1cdf3db84fff990e7e8a -s
4 S root        5263    5261    5263  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:06 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    5267  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:25 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    5268  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:04 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    5269  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:07 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    5270  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:05 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    5271  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:00 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    5275  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:05 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    5276  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:06 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    5277  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:08 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    5278  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:05 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    5279  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:08 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    5282  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:06 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    5283  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:05 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    6402  0   16  62 -18 - 590419 futex_ 06:56 ?       00:00:06 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261    7013  0   16  62 -18 - 590419 futex_ 06:57 ?       00:00:07 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5263    5261   82394  0   16  62 -18 - 590419 ep_pol 07:29 ?       00:00:03 etcd grpc-proxy start --endpoints https://192.168.126.11:9978 --metrics-addr https://0.0.0.0:9979 --listen-addr 127.0.0.1:9977 --advertise-client-url  --key /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --key-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.key --cert /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --cert-file /etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-metrics-crc.crt --cacert /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --trusted-ca-file /etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/metrics-ca-bundle.crt --listen-cipher-suites TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5273       1    5273  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/a5e3ef938e44204e0e2fd239d98d97ee026f2c255b86ab94da6f78dced593283/userdata -c a5e3ef938e44204e0e2fd239d98d97ee026f2c255b86ab94da6f78dced593283 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-etcd_etcd-crc_2139d3e2895fc6797b9c76a1b4c9886d/etcd-readyz/0.log --log-level info -n k8s_etcd-readyz_etcd-crc_openshift-etcd_2139d3e2895fc6797b9c76a1b4c9886d_0 -P /run/containers/storage/overlay-containers/a5e3ef938e44204e0e2fd239d98d97ee026f2c255b86ab94da6f78dced593283/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/a5e3ef938e44204e0e2fd239d98d97ee026f2c255b86ab94da6f78dced593283/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/a5e3ef938e44204e0e2fd239d98d97ee026f2c255b86ab94da6f78dced593283/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u a5e3ef938e44204e0e2fd239d98d97ee026f2c255b86ab94da6f78dced593283 -s
4 S root        5280    5273    5280  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:00 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    5286  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:05 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    5287  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:03 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
5 S root        5280    5273    5288  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:02 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    5289  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:03 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    5290  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:00 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    5303  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:03 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    5304  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:02 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    5305  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:03 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    5309  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:03 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    5310  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:02 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    5313  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:00 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    5314  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:02 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    5315  0   18  62 -18 - 642692 futex_ 06:56 ?       00:00:02 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    6816  0   18  62 -18 - 642692 futex_ 06:57 ?       00:00:02 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    6817  0   18  62 -18 - 642692 futex_ 06:57 ?       00:00:02 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    7564  0   18  62 -18 - 642692 futex_ 06:58 ?       00:00:00 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5280    5273    7565  0   18  62 -18 - 642692 ep_pol 06:58 ?       00:00:02 cluster-etcd-operator readyz --target=https://localhost:2379 --listen-port=9980 --serving-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.crt --serving-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-serving-crc.key --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt --listen-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256
1 S root        5292       1    5292  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/1e0670c7b36893ee38bdbacf577bd40e39cee7cfd0afcdc3a4f4f9a6f2d4096c/userdata -c 1e0670c7b36893ee38bdbacf577bd40e39cee7cfd0afcdc3a4f4f9a6f2d4096c --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-etcd_etcd-crc_2139d3e2895fc6797b9c76a1b4c9886d/etcd-rev/0.log --log-level info -n k8s_etcd-rev_etcd-crc_openshift-etcd_2139d3e2895fc6797b9c76a1b4c9886d_0 -P /run/containers/storage/overlay-containers/1e0670c7b36893ee38bdbacf577bd40e39cee7cfd0afcdc3a4f4f9a6f2d4096c/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/1e0670c7b36893ee38bdbacf577bd40e39cee7cfd0afcdc3a4f4f9a6f2d4096c/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/1e0670c7b36893ee38bdbacf577bd40e39cee7cfd0afcdc3a4f4f9a6f2d4096c/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 1e0670c7b36893ee38bdbacf577bd40e39cee7cfd0afcdc3a4f4f9a6f2d4096c -s
4 S root        5294    5292    5294  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5298  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5299  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5300  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5301  0   15  80   0 - 587137 ep_pol 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5302  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5306  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5307  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5308  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5311  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5312  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:01 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
5 S root        5294    5292    5316  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5317  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5318  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5294    5292    5319  0   15  80   0 - 587137 futex_ 06:56 ?       00:00:00 cluster-etcd-operator rev --endpoints=https://192.168.126.11:2379 --client-cert-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.crt --client-key-file=/etc/kubernetes/static-pod-certs/secrets/etcd-all-certs/etcd-peer-crc.key --client-cacert-file=/etc/kubernetes/static-pod-certs/configmaps/etcd-all-bundles/server-ca-bundle.crt
1 S root        5359       1    5359  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/d383075a74df80499c16a3151a279cf4c30167fd06de5c08ea66432c760b9f79/userdata -c d383075a74df80499c16a3151a279cf4c30167fd06de5c08ea66432c760b9f79 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log --log-level info -n k8s_kube-apiserver-check-endpoints_kube-apiserver-crc_openshift-kube-apiserver_f4b27818a5e8e43d0dc095d08835c792_1 -P /run/containers/storage/overlay-containers/d383075a74df80499c16a3151a279cf4c30167fd06de5c08ea66432c760b9f79/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/d383075a74df80499c16a3151a279cf4c30167fd06de5c08ea66432c760b9f79/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/d383075a74df80499c16a3151a279cf4c30167fd06de5c08ea66432c760b9f79/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u d383075a74df80499c16a3151a279cf4c30167fd06de5c08ea66432c760b9f79 -s
4 S root        5361    5359    5361  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5366  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:06 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5367  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
5 S root        5361    5359    5368  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:04 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5369  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5370  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5371  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:01 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5372  0   17  80   0 - 640227 ep_pol 06:56 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5373  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:03 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
5 S root        5361    5359    5374  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:04 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5375  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5401  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:04 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5403  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5404  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5405  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5406  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:02 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5361    5359    5407  0   17  80   0 - 640227 futex_ 06:56 ?       00:00:03 cluster-kube-apiserver-operator check-endpoints --kubeconfig /etc/kubernetes/static-pod-certs/configmaps/check-endpoints-kubeconfig/kubeconfig --listen 0.0.0.0:17697 --namespace openshift-kube-apiserver --v 2
1 S root        5414       1    5414  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/76030625187ffcc302718bc81c7c7a34ab47ceb0155adfe40895420a7a401c20/userdata -c 76030625187ffcc302718bc81c7c7a34ab47ceb0155adfe40895420a7a401c20 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-network-operator_iptables-alerter-4ln5h_d75a4c96-2883-4a0b-bab2-0fab2b6c0b49/iptables-alerter/0.log --log-level info -n k8s_iptables-alerter_iptables-alerter-4ln5h_openshift-network-operator_d75a4c96-2883-4a0b-bab2-0fab2b6c0b49_0 -P /run/containers/storage/overlay-containers/76030625187ffcc302718bc81c7c7a34ab47ceb0155adfe40895420a7a401c20/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/76030625187ffcc302718bc81c7c7a34ab47ceb0155adfe40895420a7a401c20/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/76030625187ffcc302718bc81c7c7a34ab47ceb0155adfe40895420a7a401c20/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 76030625187ffcc302718bc81c7c7a34ab47ceb0155adfe40895420a7a401c20 -s
1 S root        5415       1    5415  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/3dd86037c2da65d5d9cc11cd6310454968d0440cf355bd7fc22bb24838e0c05c/userdata -c 3dd86037c2da65d5d9cc11cd6310454968d0440cf355bd7fc22bb24838e0c05c --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-network-node-identity_network-node-identity-vrzqb_ef543e1b-8068-4ea3-b32a-61027b32e95d/webhook/0.log --log-level info -n k8s_webhook_network-node-identity-vrzqb_openshift-network-node-identity_ef543e1b-8068-4ea3-b32a-61027b32e95d_0 -P /run/containers/storage/overlay-containers/3dd86037c2da65d5d9cc11cd6310454968d0440cf355bd7fc22bb24838e0c05c/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/3dd86037c2da65d5d9cc11cd6310454968d0440cf355bd7fc22bb24838e0c05c/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/3dd86037c2da65d5d9cc11cd6310454968d0440cf355bd7fc22bb24838e0c05c/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 3dd86037c2da65d5d9cc11cd6310454968d0440cf355bd7fc22bb24838e0c05c -s
4 S 1000470+    5417    5415    5417  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    5450  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    5451  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    5453  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    5455  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    5457  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    5469  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    5473  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    5474  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    5475  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    5476  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    5956  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    5957  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    6133  0   18  80   0 - 634670 ep_pol 06:56 ?       00:00:01 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    6134  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415    6135  0   18  80   0 - 634670 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415   11665  0   18  80   0 - 634670 futex_ 06:58 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S 1000470+    5417    5415   85095  0   18  80   0 - 634670 futex_ 07:31 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --webhook-cert-dir=/etc/webhook-cert --webhook-host=127.0.0.1 --webhook-port=9743 --enable-hybrid-overlay --enable-interconnect --disable-approver --extra-allowed-user=system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane --wait-for-kubernetes-api=200s --pod-admission-conditions=/var/run/ovnkube-identity-config/additional-pod-admission-cond.json --loglevel=2
1 S root        5421       1    5421  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/49c5a7168b89d25e54930685c65c87b923f07da4c2661309a20be99aae4226a9/userdata -c 49c5a7168b89d25e54930685c65c87b923f07da4c2661309a20be99aae4226a9 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-network-operator_network-operator-58b4c7f79c-55gtf_37a5e44f-9a88-4405-be8a-b645485e7312/network-operator/0.log --log-level info -n k8s_network-operator_network-operator-58b4c7f79c-55gtf_openshift-network-operator_37a5e44f-9a88-4405-be8a-b645485e7312_0 -P /run/containers/storage/overlay-containers/49c5a7168b89d25e54930685c65c87b923f07da4c2661309a20be99aae4226a9/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/49c5a7168b89d25e54930685c65c87b923f07da4c2661309a20be99aae4226a9/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/49c5a7168b89d25e54930685c65c87b923f07da4c2661309a20be99aae4226a9/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 49c5a7168b89d25e54930685c65c87b923f07da4c2661309a20be99aae4226a9 -s
4 S nfsnobo+    5423    5421    5423  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:00 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
1 R nfsnobo+    5423    5421    5459  0   18  80   0 - 675881 -     06:56 ?        00:00:15 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
5 S nfsnobo+    5423    5421    5460  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:00 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
5 S nfsnobo+    5423    5421    5461  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:06 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
5 S nfsnobo+    5423    5421    5462  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:06 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
1 S nfsnobo+    5423    5421    5463  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:00 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
1 S nfsnobo+    5423    5421    5470  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:04 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
1 S nfsnobo+    5423    5421    5471  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:00 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
1 S nfsnobo+    5423    5421    5472  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:01 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
5 S nfsnobo+    5423    5421    5479  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:01 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
1 S nfsnobo+    5423    5421    5548  0   18  80   0 - 675881 ep_pol 06:56 ?       00:00:03 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
1 S nfsnobo+    5423    5421    5549  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:00 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
1 S nfsnobo+    5423    5421    5550  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:07 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
1 S nfsnobo+    5423    5421    5551  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:06 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
1 S nfsnobo+    5423    5421    5581  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:07 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
1 S nfsnobo+    5423    5421    5582  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:03 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
5 S nfsnobo+    5423    5421    5585  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:02 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
5 S nfsnobo+    5423    5421    6062  0   18  80   0 - 675881 futex_ 06:56 ?       00:00:03 /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104
1 S root        5432       1    5432  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/9299b7c0a1edfff07f1a3abcaa131156b070d58980b9ca81f1b5b13ba0a64e24/userdata -c 9299b7c0a1edfff07f1a3abcaa131156b070d58980b9ca81f1b5b13ba0a64e24 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-dns_node-resolver-d69n4_4d67ab19-ac19-4673-ade0-a35ffb299e85/dns-node-resolver/0.log --log-level info -n k8s_dns-node-resolver_node-resolver-d69n4_openshift-dns_4d67ab19-ac19-4673-ade0-a35ffb299e85_0 -P /run/containers/storage/overlay-containers/9299b7c0a1edfff07f1a3abcaa131156b070d58980b9ca81f1b5b13ba0a64e24/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/9299b7c0a1edfff07f1a3abcaa131156b070d58980b9ca81f1b5b13ba0a64e24/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/9299b7c0a1edfff07f1a3abcaa131156b070d58980b9ca81f1b5b13ba0a64e24/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 9299b7c0a1edfff07f1a3abcaa131156b070d58980b9ca81f1b5b13ba0a64e24 -s
1 S root        5434       1    5434  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/2c1cb407e6a1c1da2286d3c36866dffac7b42b07230395e0cc77c56d8ac29629/userdata -c 2c1cb407e6a1c1da2286d3c36866dffac7b42b07230395e0cc77c56d8ac29629 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-network-node-identity_network-node-identity-vrzqb_ef543e1b-8068-4ea3-b32a-61027b32e95d/approver/0.log --log-level info -n k8s_approver_network-node-identity-vrzqb_openshift-network-node-identity_ef543e1b-8068-4ea3-b32a-61027b32e95d_0 -P /run/containers/storage/overlay-containers/2c1cb407e6a1c1da2286d3c36866dffac7b42b07230395e0cc77c56d8ac29629/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/2c1cb407e6a1c1da2286d3c36866dffac7b42b07230395e0cc77c56d8ac29629/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/2c1cb407e6a1c1da2286d3c36866dffac7b42b07230395e0cc77c56d8ac29629/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 2c1cb407e6a1c1da2286d3c36866dffac7b42b07230395e0cc77c56d8ac29629 -s
4 S root        5436    5432    5436  0    1  80   0 -  1208 do_wai 06:56 ?        00:00:00 /bin/bash -c #!/bin/bash set -uo pipefail  trap 'jobs -p | xargs kill || true; wait; exit 0' TERM  OPENSHIFT_MARKER="openshift-generated-node-resolver" HOSTS_FILE="/etc/hosts" TEMP_FILE="/etc/hosts.tmp"  IFS=', ' read -r -a services <<< "${SERVICES}"  # Make a temporary file with the old hosts file's attributes. if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then   echo "Failed to preserve hosts file. Exiting."   exit 1 fi  while true; do   declare -A svc_ips   for svc in "${services[@]}"; do     # Fetch service IP from cluster dns if present. We make several tries     # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones     # are for deployments with Kuryr on older OpenStack (OSP13) - those do not     # support UDP loadbalancers and require reaching DNS through TCP.     cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"'           'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"'           'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"'           'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"')     for i in ${!cmds[*]}     do       ips=($(eval "${cmds[i]}"))       if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then         svc_ips["${svc}"]="${ips[@]}"         break       fi     done   done    # Update /etc/hosts only if we get valid service IPs   # We will not update /etc/hosts when there is coredns service outage or api unavailability   # Stale entries could exist in /etc/hosts if the service is deleted   if [[ -n "${svc_ips[*]-}" ]]; then     # Build a new hosts file from /etc/hosts with our custom entries filtered out     if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then       # Only continue rebuilding the hosts entries if its original content is preserved       sleep 60 & wait       continue     fi      # Append resolver entries for services     rc=0     for svc in "${!svc_ips[@]}"; do       for ip in ${svc_ips[${svc}]}; do         echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$?       done     done     if [[ $rc -ne 0 ]]; then       sleep 60 & wait       continue     fi       # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior     # Replace /etc/hosts with our modified version if needed     cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}"     # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn   fi   sleep 60 & wait   unset svc_ips done 
4 S 1000470+    5437    5434    5437  0   12  80   0 - 523560 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --disable-webhook --csr-acceptance-conditions=/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json --loglevel=4
1 S 1000470+    5437    5434    5449  0   12  80   0 - 523560 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --disable-webhook --csr-acceptance-conditions=/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json --loglevel=4
1 S 1000470+    5437    5434    5452  0   12  80   0 - 523560 ep_pol 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --disable-webhook --csr-acceptance-conditions=/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json --loglevel=4
5 S 1000470+    5437    5434    5454  0   12  80   0 - 523560 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --disable-webhook --csr-acceptance-conditions=/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json --loglevel=4
1 S 1000470+    5437    5434    5456  0   12  80   0 - 523560 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --disable-webhook --csr-acceptance-conditions=/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json --loglevel=4
1 S 1000470+    5437    5434    5458  0   12  80   0 - 523560 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --disable-webhook --csr-acceptance-conditions=/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json --loglevel=4
1 S 1000470+    5437    5434    5468  0   12  80   0 - 523560 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --disable-webhook --csr-acceptance-conditions=/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json --loglevel=4
1 S 1000470+    5437    5434    5477  0   12  80   0 - 523560 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --disable-webhook --csr-acceptance-conditions=/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json --loglevel=4
1 S 1000470+    5437    5434    5478  0   12  80   0 - 523560 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --disable-webhook --csr-acceptance-conditions=/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json --loglevel=4
1 S 1000470+    5437    5434    5480  0   12  80   0 - 523560 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --disable-webhook --csr-acceptance-conditions=/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json --loglevel=4
1 S 1000470+    5437    5434    5481  0   12  80   0 - 523560 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --disable-webhook --csr-acceptance-conditions=/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json --loglevel=4
1 S 1000470+    5437    5434    5482  0   12  80   0 - 523560 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 --disable-webhook --csr-acceptance-conditions=/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json --loglevel=4
4 S root        5483    5414    5483  0    1  80   0 -  1080 do_wai 06:56 ?        00:00:00 /bin/bash /iptables-alerter/iptables-alerter.sh
1 S root        5552       1    5552  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/3cce49d494ea2a4174add0a3f5532eca27afe61c7fa807f6e25c594efe4bee7e/userdata -c 3cce49d494ea2a4174add0a3f5532eca27afe61c7fa807f6e25c594efe4bee7e --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-machine-config-operator_machine-config-daemon-z8f25_e12d3cfb-2ba7-4eb6-b6b4-bfc4cec93930/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_machine-config-daemon-z8f25_openshift-machine-config-operator_e12d3cfb-2ba7-4eb6-b6b4-bfc4cec93930_0 -P /run/containers/storage/overlay-containers/3cce49d494ea2a4174add0a3f5532eca27afe61c7fa807f6e25c594efe4bee7e/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/3cce49d494ea2a4174add0a3f5532eca27afe61c7fa807f6e25c594efe4bee7e/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/3cce49d494ea2a4174add0a3f5532eca27afe61c7fa807f6e25c594efe4bee7e/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 3cce49d494ea2a4174add0a3f5532eca27afe61c7fa807f6e25c594efe4bee7e -s
4 S nfsnobo+    5557    5552    5557  0   13  80   0 - 540855 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    5557    5552    5563  0   13  80   0 - 540855 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    5557    5552    5564  0   13  80   0 - 540855 ep_pol 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    5557    5552    5565  0   13  80   0 - 540855 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    5557    5552    5566  0   13  80   0 - 540855 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    5557    5552    5567  0   13  80   0 - 540855 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    5557    5552    5568  0   13  80   0 - 540855 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    5557    5552    5569  0   13  80   0 - 540855 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    5557    5552    5570  0   13  80   0 - 540855 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    5557    5552    5571  0   13  80   0 - 540855 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    5557    5552    5572  0   13  80   0 - 540855 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    5557    5552    5573  0   13  80   0 - 540855 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    5557    5552    5574  0   13  80   0 - 540855 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S root        5805       1    5805  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/bd8ce586c6c3ce067f3d73acbee488921afe3f125b0d750e98ee225b5a04690a/userdata -c bd8ce586c6c3ce067f3d73acbee488921afe3f125b0d750e98ee225b5a04690a --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-image-registry_node-ca-vnb5p_59e2d65e-f3c1-4b45-8c19-b033bd5f3aac/node-ca/0.log --log-level info -n k8s_node-ca_node-ca-vnb5p_openshift-image-registry_59e2d65e-f3c1-4b45-8c19-b033bd5f3aac_0 -P /run/containers/storage/overlay-containers/bd8ce586c6c3ce067f3d73acbee488921afe3f125b0d750e98ee225b5a04690a/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/bd8ce586c6c3ce067f3d73acbee488921afe3f125b0d750e98ee225b5a04690a/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/bd8ce586c6c3ce067f3d73acbee488921afe3f125b0d750e98ee225b5a04690a/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u bd8ce586c6c3ce067f3d73acbee488921afe3f125b0d750e98ee225b5a04690a -s
4 S 1001        5809    5805    5809  0    1  80   0 -  1176 do_wai 06:56 ?        00:00:00 /bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM while [ true ]; do   for f in $(ls /tmp/serviceca); do       echo $f       ca_file_path="/tmp/serviceca/${f}"       f=$(echo $f | sed  -r 's/(.*)\.\./\1:/')       reg_dir_path="/etc/docker/certs.d/${f}"       if [ -e "${reg_dir_path}" ]; then           cp -u $ca_file_path $reg_dir_path/ca.crt       else           mkdir $reg_dir_path           cp $ca_file_path $reg_dir_path/ca.crt       fi   done   for d in $(ls /etc/docker/certs.d); do       echo $d       dp=$(echo $d | sed  -r 's/(.*):/\1\.\./')       reg_conf_path="/tmp/serviceca/${dp}"       if [ ! -e "${reg_conf_path}" ]; then           rm -rf /etc/docker/certs.d/$d       fi   done   sleep 60 & wait ${!} done 
1 S root        6151       1    6151  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/dccb87a733522fc4051eb563645de5cf31771f112be5c8129e5a9401964d4574/userdata -c dccb87a733522fc4051eb563645de5cf31771f112be5c8129e5a9401964d4574 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-multus_multus-additional-cni-plugins-6fdnr_8a71d33c-dc75-4c28-bda0-0b3793de7de8/kube-multus-additional-cni-plugins/0.log --log-level info -n k8s_kube-multus-additional-cni-plugins_multus-additional-cni-plugins-6fdnr_openshift-multus_8a71d33c-dc75-4c28-bda0-0b3793de7de8_0 -P /run/containers/storage/overlay-containers/dccb87a733522fc4051eb563645de5cf31771f112be5c8129e5a9401964d4574/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/dccb87a733522fc4051eb563645de5cf31771f112be5c8129e5a9401964d4574/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/dccb87a733522fc4051eb563645de5cf31771f112be5c8129e5a9401964d4574/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u dccb87a733522fc4051eb563645de5cf31771f112be5c8129e5a9401964d4574 -s
4 S root        6153    6151    6153  0    1  80   0 -  1047 do_wai 06:56 ?        00:00:00 /bin/bash -ec -- trap : TERM INT; sleep infinity & wait 
4 S root        6157    6153    6157  0    1  80   0 -  1233 do_sys 06:56 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep infinity
1 S root        6330       1    6330  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/249c0497df0b8af6b7a68e0b6af6cad0a9f6552e36820e2c1e9734ab7cc04f9d/userdata -c 249c0497df0b8af6b7a68e0b6af6cad0a9f6552e36820e2c1e9734ab7cc04f9d --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ovn-kubernetes_ovnkube-control-plane-749d76644c-pz9mk_44db27b9-37f2-443a-8c72-3109af8b80bf/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_ovnkube-control-plane-749d76644c-pz9mk_openshift-ovn-kubernetes_44db27b9-37f2-443a-8c72-3109af8b80bf_0 -P /run/containers/storage/overlay-containers/249c0497df0b8af6b7a68e0b6af6cad0a9f6552e36820e2c1e9734ab7cc04f9d/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/249c0497df0b8af6b7a68e0b6af6cad0a9f6552e36820e2c1e9734ab7cc04f9d/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/249c0497df0b8af6b7a68e0b6af6cad0a9f6552e36820e2c1e9734ab7cc04f9d/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 249c0497df0b8af6b7a68e0b6af6cad0a9f6552e36820e2c1e9734ab7cc04f9d -s
4 S nfsnobo+    6332    6330    6332  0   13  80   0 - 540791 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+    6332    6330    6338  0   13  80   0 - 540791 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+    6332    6330    6339  0   13  80   0 - 540791 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+    6332    6330    6340  0   13  80   0 - 540791 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+    6332    6330    6341  0   13  80   0 - 540791 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+    6332    6330    6342  0   13  80   0 - 540791 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+    6332    6330    6344  0   13  80   0 - 540791 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+    6332    6330    6345  0   13  80   0 - 540791 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+    6332    6330    6346  0   13  80   0 - 540791 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+    6332    6330    6348  0   13  80   0 - 540791 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+    6332    6330    6349  0   13  80   0 - 540791 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+    6332    6330    6350  0   13  80   0 - 540791 futex_ 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+    6332    6330    6351  0   13  80   0 - 540791 ep_pol 06:56 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9108 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29108/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S root        6352       1    6352  0    1  80   0 -  2078 do_pol 06:56 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/b93301c77ee354b61202d8e145fde12e30681b03784fee544edf0cc915306d0a/userdata -c b93301c77ee354b61202d8e145fde12e30681b03784fee544edf0cc915306d0a --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ovn-kubernetes_ovnkube-control-plane-749d76644c-pz9mk_44db27b9-37f2-443a-8c72-3109af8b80bf/ovnkube-cluster-manager/0.log --log-level info -n k8s_ovnkube-cluster-manager_ovnkube-control-plane-749d76644c-pz9mk_openshift-ovn-kubernetes_44db27b9-37f2-443a-8c72-3109af8b80bf_0 -P /run/containers/storage/overlay-containers/b93301c77ee354b61202d8e145fde12e30681b03784fee544edf0cc915306d0a/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/b93301c77ee354b61202d8e145fde12e30681b03784fee544edf0cc915306d0a/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/b93301c77ee354b61202d8e145fde12e30681b03784fee544edf0cc915306d0a/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u b93301c77ee354b61202d8e145fde12e30681b03784fee544edf0cc915306d0a -s
4 S root        6354    6352    6354  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6359  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:03 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6360  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6361  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6362  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6363  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6365  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6366  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6367  0   17  80   0 - 636455 ep_pol 06:56 ?       00:00:03 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6368  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
5 S root        6354    6352    6369  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:03 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6370  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6371  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6372  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6373  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:02 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6374  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:00 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        6354    6352    6375  0   17  80   0 - 636455 futex_ 06:56 ?       00:00:01 /usr/bin/ovnkube --enable-interconnect --init-cluster-manager crc --config-file=/run/ovnkube-config/ovnkube.conf --loglevel 4 --metrics-bind-address 127.0.0.1:29108 --metrics-enable-pprof --metrics-enable-config-duration --enable-persistent-ips --enable-multi-network --enable-network-segmentation
1 S root        7126       1    7126  0    1  80   0 -  2078 do_pol 06:57 ?        00:00:03 /usr/bin/conmon -b /run/containers/storage/overlay-containers/47e39444be3201d3d70eaff74a336c83ce80abc6f10c8cda1e49709bf404ca67/userdata -c 47e39444be3201d3d70eaff74a336c83ce80abc6f10c8cda1e49709bf404ca67 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-cluster-version_cluster-version-operator-5c965bbfc6-s75r8_944b28dd-5656-4765-aa44-a527384e9188/cluster-version-operator/0.log --log-level info -n k8s_cluster-version-operator_cluster-version-operator-5c965bbfc6-s75r8_openshift-cluster-version_944b28dd-5656-4765-aa44-a527384e9188_0 -P /run/containers/storage/overlay-containers/47e39444be3201d3d70eaff74a336c83ce80abc6f10c8cda1e49709bf404ca67/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/47e39444be3201d3d70eaff74a336c83ce80abc6f10c8cda1e49709bf404ca67/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/47e39444be3201d3d70eaff74a336c83ce80abc6f10c8cda1e49709bf404ca67/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 47e39444be3201d3d70eaff74a336c83ce80abc6f10c8cda1e49709bf404ca67 -s
4 S 1000040+    7128    7126    7128  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:00 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
1 S 1000040+    7128    7126    7132  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:10 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7133  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:07 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7134  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:06 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
1 S 1000040+    7128    7126    7135  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:00 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
1 S 1000040+    7128    7126    7136  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:00 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7137  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:06 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7138  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:06 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7139  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:07 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7140  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:05 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7141  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:06 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7142  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:05 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
1 S 1000040+    7128    7126    7143  0   19  80   0 - 675363 ep_pol 06:57 ?       00:00:00 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7144  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:06 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7145  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:05 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7146  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:06 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7147  0   19  80   0 - 675363 ep_pol 06:57 ?       00:00:05 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
5 S 1000040+    7128    7126    7148  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:07 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
1 S 1000040+    7128    7126    7207  0   19  80   0 - 675363 futex_ 06:57 ?       00:00:07 /usr/bin/cluster-version-operator start --release-image=quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03 --enable-auto-update=false --listen=0.0.0.0:9099 --serving-cert-file=/etc/tls/serving-cert/tls.crt --serving-key-file=/etc/tls/serving-cert/tls.key --v=2 --always-enable-capabilities=Ingress
1 S root        7756       1    7756  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/b3908f9383ca63feffafc9259e5bac7736256289a7e1a2e4db041397027d8dd8/userdata -c b3908f9383ca63feffafc9259e5bac7736256289a7e1a2e4db041397027d8dd8 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ingress-operator_ingress-operator-5b745b69d9-8mqfc_f02cd7f0-75f4-4c17-8726-7a592a95d6f9/ingress-operator/0.log --log-level info -n k8s_ingress-operator_ingress-operator-5b745b69d9-8mqfc_openshift-ingress-operator_f02cd7f0-75f4-4c17-8726-7a592a95d6f9_0 -P /run/containers/storage/overlay-containers/b3908f9383ca63feffafc9259e5bac7736256289a7e1a2e4db041397027d8dd8/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/b3908f9383ca63feffafc9259e5bac7736256289a7e1a2e4db041397027d8dd8/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/b3908f9383ca63feffafc9259e5bac7736256289a7e1a2e4db041397027d8dd8/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u b3908f9383ca63feffafc9259e5bac7736256289a7e1a2e4db041397027d8dd8 -s
4 S 1000120+    7769    7756    7769  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:00 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8119  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:08 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8122  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:04 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8124  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:00 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8125  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:00 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8134  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:00 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8552  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:03 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8553  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:03 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8556  0   18  80   0 - 676967 ep_pol 06:58 ?       00:00:00 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8666  0   18  80   0 - 676967 ep_pol 06:58 ?       00:00:00 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
5 S 1000120+    7769    7756    8881  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:02 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8909  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:04 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8911  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:01 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8912  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:00 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8913  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:00 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8916  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:00 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8917  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:02 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S 1000120+    7769    7756    8961  0   18  80   0 - 676967 futex_ 06:58 ?       00:00:02 ingress-operator start --namespace openshift-ingress-operator --image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b --canary-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e --release-version 4.18.1
1 S root        8062       1    8062  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/16c1b31ae29308332df87c8d7b2edc27385ccc3b031e773122fe8b29113b8ad0/userdata -c 16c1b31ae29308332df87c8d7b2edc27385ccc3b031e773122fe8b29113b8ad0 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ingress-operator_ingress-operator-5b745b69d9-8mqfc_f02cd7f0-75f4-4c17-8726-7a592a95d6f9/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_ingress-operator-5b745b69d9-8mqfc_openshift-ingress-operator_f02cd7f0-75f4-4c17-8726-7a592a95d6f9_0 -P /run/containers/storage/overlay-containers/16c1b31ae29308332df87c8d7b2edc27385ccc3b031e773122fe8b29113b8ad0/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/16c1b31ae29308332df87c8d7b2edc27385ccc3b031e773122fe8b29113b8ad0/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/16c1b31ae29308332df87c8d7b2edc27385ccc3b031e773122fe8b29113b8ad0/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 16c1b31ae29308332df87c8d7b2edc27385ccc3b031e773122fe8b29113b8ad0 -s
4 S 1000120+    8073    8062    8073  0   10  80   0 - 485364 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000120+    8073    8062    8212  0   10  80   0 - 485364 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000120+    8073    8062    8213  0   10  80   0 - 485364 ep_pol 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000120+    8073    8062    8214  0   10  80   0 - 485364 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000120+    8073    8062    8215  0   10  80   0 - 485364 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000120+    8073    8062    8216  0   10  80   0 - 485364 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000120+    8073    8062    8220  0   10  80   0 - 485364 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000120+    8073    8062    8223  0   10  80   0 - 485364 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000120+    8073    8062    8224  0   10  80   0 - 485364 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000120+    8073    8062   17223  0   10  80   0 - 485364 futex_ 07:00 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S root        8316       1    8316  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/9b327fa9fea589ad085d1dc386b87d1e3683b76c7b012e7d494607d83cf2c07f/userdata -c 9b327fa9fea589ad085d1dc386b87d1e3683b76c7b012e7d494607d83cf2c07f --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-apiserver-operator_openshift-apiserver-operator-796bbdcf4f-782lq_9a499f8e-4807-4432-810e-b240eae2b261/openshift-apiserver-operator/0.log --log-level info -n k8s_openshift-apiserver-operator_openshift-apiserver-operator-796bbdcf4f-782lq_openshift-apiserver-operator_9a499f8e-4807-4432-810e-b240eae2b261_0 -P /run/containers/storage/overlay-containers/9b327fa9fea589ad085d1dc386b87d1e3683b76c7b012e7d494607d83cf2c07f/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/9b327fa9fea589ad085d1dc386b87d1e3683b76c7b012e7d494607d83cf2c07f/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/9b327fa9fea589ad085d1dc386b87d1e3683b76c7b012e7d494607d83cf2c07f/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 9b327fa9fea589ad085d1dc386b87d1e3683b76c7b012e7d494607d83cf2c07f -s
4 S nfsnobo+    8322    8316    8322  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:00 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    8618  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:15 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    8619  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:05 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    8620  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:05 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    8621  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:05 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    8622  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:00 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    8762  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:05 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    8891  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:05 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    8892  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:05 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
5 S nfsnobo+    8322    8316    8972  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:01 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    8973  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:00 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    9433  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:05 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    9436  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:06 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    9438  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:05 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    9439  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:05 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316    9440  0   18  80   0 - 675238 ep_pol 06:58 ?       00:00:06 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316   10569  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:05 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8322    8316   10570  0   18  80   0 - 675238 futex_ 06:58 ?       00:00:06 cluster-openshift-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S root        8479       1    8479  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/ba1bfc2a8cdb96ab126d7d81bae516e800599fe4cf55c554dc044dd81416b9b2/userdata -c ba1bfc2a8cdb96ab126d7d81bae516e800599fe4cf55c554dc044dd81416b9b2 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-console-operator_console-operator-58897d9998-cn2tr_ec569cec-4aef-4da7-8c8b-9a14f5565471/console-operator/0.log --log-level info -n k8s_console-operator_console-operator-58897d9998-cn2tr_openshift-console-operator_ec569cec-4aef-4da7-8c8b-9a14f5565471_0 -P /run/containers/storage/overlay-containers/ba1bfc2a8cdb96ab126d7d81bae516e800599fe4cf55c554dc044dd81416b9b2/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/ba1bfc2a8cdb96ab126d7d81bae516e800599fe4cf55c554dc044dd81416b9b2/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/ba1bfc2a8cdb96ab126d7d81bae516e800599fe4cf55c554dc044dd81416b9b2/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u ba1bfc2a8cdb96ab126d7d81bae516e800599fe4cf55c554dc044dd81416b9b2 -s
4 S 1000610+    8489    8479    8489  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:00 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479    8658  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:11 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479    8659  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:03 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479    8660  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:03 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479    8661  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:00 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479    8663  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:00 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479    8914  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:05 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479    8962  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:03 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479    9211  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:03 console operator --config=/var/run/configmaps/config/controller-config.yaml
5 S 1000610+    8489    8479    9212  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:04 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479    9333  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:03 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479    9334  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:04 console operator --config=/var/run/configmaps/config/controller-config.yaml
5 S 1000610+    8489    8479    9335  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:04 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479    9336  0   17  80   0 - 623337 ep_pol 06:58 ?       00:00:03 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479    9530  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:02 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479   11529  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:03 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S 1000610+    8489    8479   11530  0   17  80   0 - 623337 futex_ 06:58 ?       00:00:03 console operator --config=/var/run/configmaps/config/controller-config.yaml
1 S root        8645       1    8645  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/9139dc89bfa74882870bc0a12debc139f8946084d45006ba4d221646c52c4ac5/userdata -c 9139dc89bfa74882870bc0a12debc139f8946084d45006ba4d221646c52c4ac5 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-cluster-machine-approver_machine-approver-56656f9798-f5ggg_db163501-b4f0-48b4-a558-e7f3d9d1c835/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_machine-approver-56656f9798-f5ggg_openshift-cluster-machine-approver_db163501-b4f0-48b4-a558-e7f3d9d1c835_0 -P /run/containers/storage/overlay-containers/9139dc89bfa74882870bc0a12debc139f8946084d45006ba4d221646c52c4ac5/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/9139dc89bfa74882870bc0a12debc139f8946084d45006ba4d221646c52c4ac5/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/9139dc89bfa74882870bc0a12debc139f8946084d45006ba4d221646c52c4ac5/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 9139dc89bfa74882870bc0a12debc139f8946084d45006ba4d221646c52c4ac5 -s
1 S root        8646       1    8646  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/009e4702f2ec77011db2258d9207f164c63fe88fb716d494a1c552d8576c5625/userdata -c 009e4702f2ec77011db2258d9207f164c63fe88fb716d494a1c552d8576c5625 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-zt5t2_d131a4aa-4055-4aa9-bfef-4354654e6577/009e4702f2ec77011db2258d9207f164c63fe88fb716d494a1c552d8576c5625.log --log-level info -n k8s_POD_cluster-samples-operator-665b6dd947-zt5t2_openshift-cluster-samples-operator_d131a4aa-4055-4aa9-bfef-4354654e6577_0 -P /run/containers/storage/overlay-containers/009e4702f2ec77011db2258d9207f164c63fe88fb716d494a1c552d8576c5625/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/009e4702f2ec77011db2258d9207f164c63fe88fb716d494a1c552d8576c5625/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/009e4702f2ec77011db2258d9207f164c63fe88fb716d494a1c552d8576c5625/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 009e4702f2ec77011db2258d9207f164c63fe88fb716d494a1c552d8576c5625 -s
4 S nfsnobo+    8654    8645    8654  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9192 --upstream=http://127.0.0.1:9191/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S nfsnobo+    8654    8645    8927  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9192 --upstream=http://127.0.0.1:9191/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S nfsnobo+    8654    8645    8928  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9192 --upstream=http://127.0.0.1:9191/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S nfsnobo+    8654    8645    8929  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9192 --upstream=http://127.0.0.1:9191/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S nfsnobo+    8654    8645    8930  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9192 --upstream=http://127.0.0.1:9191/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S nfsnobo+    8654    8645    8931  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9192 --upstream=http://127.0.0.1:9191/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S nfsnobo+    8654    8645    8934  0   11  80   0 - 503733 ep_pol 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9192 --upstream=http://127.0.0.1:9191/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S nfsnobo+    8654    8645    8968  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9192 --upstream=http://127.0.0.1:9191/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S nfsnobo+    8654    8645    8969  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9192 --upstream=http://127.0.0.1:9191/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S nfsnobo+    8654    8645   17225  0   11  80   0 - 503733 futex_ 07:00 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9192 --upstream=http://127.0.0.1:9191/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S nfsnobo+    8654    8645  131063  0   11  80   0 - 503733 futex_ 08:15 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9192 --upstream=http://127.0.0.1:9191/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
4 S root        8655    8646    8655  0    1  80   0 -   624 do_sys 06:58 ?        00:00:00 /usr/bin/pod
1 S root        8937       1    8937  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/350d4f7f069e71e47beaee782fe17b2f87b6f48d162d3aca463b67c108e92e70/userdata -c 350d4f7f069e71e47beaee782fe17b2f87b6f48d162d3aca463b67c108e92e70 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-controller-manager-operator_openshift-controller-manager-operator-756b6f6bc6-x6rz8_17e82ce0-83fd-416a-aab4-4a20524f89d4/openshift-controller-manager-operator/0.log --log-level info -n k8s_openshift-controller-manager-operator_openshift-controller-manager-operator-756b6f6bc6-x6rz8_openshift-controller-manager-operator_17e82ce0-83fd-416a-aab4-4a20524f89d4_0 -P /run/containers/storage/overlay-containers/350d4f7f069e71e47beaee782fe17b2f87b6f48d162d3aca463b67c108e92e70/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/350d4f7f069e71e47beaee782fe17b2f87b6f48d162d3aca463b67c108e92e70/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/350d4f7f069e71e47beaee782fe17b2f87b6f48d162d3aca463b67c108e92e70/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 350d4f7f069e71e47beaee782fe17b2f87b6f48d162d3aca463b67c108e92e70 -s
1 S root        8939       1    8939  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/5aa3f8d4144498ffac9d4e2a869636e5a96c5b005244ad0b2e3971faca99cd64/userdata -c 5aa3f8d4144498ffac9d4e2a869636e5a96c5b005244ad0b2e3971faca99cd64 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-image-registry_cluster-image-registry-operator-dc59b4c8b-xqmbj_f2031c6f-7934-4ab7-aefb-6c11c8fcb48f/cluster-image-registry-operator/0.log --log-level info -n k8s_cluster-image-registry-operator_cluster-image-registry-operator-dc59b4c8b-xqmbj_openshift-image-registry_f2031c6f-7934-4ab7-aefb-6c11c8fcb48f_0 -P /run/containers/storage/overlay-containers/5aa3f8d4144498ffac9d4e2a869636e5a96c5b005244ad0b2e3971faca99cd64/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/5aa3f8d4144498ffac9d4e2a869636e5a96c5b005244ad0b2e3971faca99cd64/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/5aa3f8d4144498ffac9d4e2a869636e5a96c5b005244ad0b2e3971faca99cd64/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 5aa3f8d4144498ffac9d4e2a869636e5a96c5b005244ad0b2e3971faca99cd64 -s
1 S root        8942       1    8942  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/8dd9811ad9a018a5954b4d99e88ad4e6e43b9ac28200236667ee5073c0109a45/userdata -c 8dd9811ad9a018a5954b4d99e88ad4e6e43b9ac28200236667ee5073c0109a45 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ingress_router-default-5444994796-xt6fn_0710cdc3-8aa1-4d3b-8ab2-1ff402b20941/router/0.log --log-level info -n k8s_router_router-default-5444994796-xt6fn_openshift-ingress_0710cdc3-8aa1-4d3b-8ab2-1ff402b20941_0 -P /run/containers/storage/overlay-containers/8dd9811ad9a018a5954b4d99e88ad4e6e43b9ac28200236667ee5073c0109a45/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/8dd9811ad9a018a5954b4d99e88ad4e6e43b9ac28200236667ee5073c0109a45/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/8dd9811ad9a018a5954b4d99e88ad4e6e43b9ac28200236667ee5073c0109a45/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 8dd9811ad9a018a5954b4d99e88ad4e6e43b9ac28200236667ee5073c0109a45 -s
4 S nfsnobo+    8947    8937    8947  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:00 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9163  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:14 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
5 S nfsnobo+    8947    8937    9166  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:01 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9168  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:02 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9169  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:04 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9179  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:00 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9492  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:00 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9495  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:06 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9563  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:02 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9716  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:04 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9717  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:05 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9718  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:03 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9953  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:01 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9955  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:03 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9957  0   17  80   0 - 639265 ep_pol 06:58 ?       00:00:05 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937    9973  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:02 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    8947    8937   10007  0   17  80   0 - 639265 futex_ 06:58 ?       00:00:04 cluster-openshift-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
4 S 1000560+    8949    8942    8949  0   13  80   0 - 542678 futex_ 06:58 ?       00:00:00 /usr/bin/openshift-router --v=2
1 S 1000560+    8949    8942    9164  0   13  80   0 - 542678 futex_ 06:58 ?       00:00:01 /usr/bin/openshift-router --v=2
5 S 1000560+    8949    8942    9170  0   13  80   0 - 542678 futex_ 06:58 ?       00:00:00 /usr/bin/openshift-router --v=2
5 S 1000560+    8949    8942    9171  0   13  80   0 - 542678 futex_ 06:58 ?       00:00:00 /usr/bin/openshift-router --v=2
1 S 1000560+    8949    8942    9172  0   13  80   0 - 542678 futex_ 06:58 ?       00:00:00 /usr/bin/openshift-router --v=2
5 S 1000560+    8949    8942    9173  0   13  80   0 - 542678 ep_pol 06:58 ?       00:00:01 /usr/bin/openshift-router --v=2
1 S 1000560+    8949    8942    9181  0   13  80   0 - 542678 futex_ 06:58 ?       00:00:00 /usr/bin/openshift-router --v=2
5 S 1000560+    8949    8942    9351  0   13  80   0 - 542678 futex_ 06:58 ?       00:00:01 /usr/bin/openshift-router --v=2
5 S 1000560+    8949    8942    9435  0   13  80   0 - 542678 futex_ 06:58 ?       00:00:00 /usr/bin/openshift-router --v=2
5 S 1000560+    8949    8942    9443  0   13  80   0 - 542678 futex_ 06:58 ?       00:00:00 /usr/bin/openshift-router --v=2
5 S 1000560+    8949    8942   15301  0   13  80   0 - 542678 futex_ 06:59 ?       00:00:01 /usr/bin/openshift-router --v=2
5 S 1000560+    8949    8942   15302  0   13  80   0 - 542678 futex_ 06:59 ?       00:00:00 /usr/bin/openshift-router --v=2
5 S 1000560+    8949    8942   15303  0   13  80   0 - 542678 futex_ 06:59 ?       00:00:01 /usr/bin/openshift-router --v=2
4 S 1000260+    8952    8939    8952  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:00 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9206  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:05 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9207  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:00 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
5 S 1000260+    8952    8939    9208  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:02 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9210  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:01 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9224  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:00 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9373  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:01 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9374  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:00 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9375  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:02 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9426  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:00 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9818  0   18  80   0 - 646405 ep_pol 06:58 ?       00:00:01 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9828  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:01 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
5 S 1000260+    8952    8939    9829  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:03 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9830  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:00 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9832  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:02 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9881  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:00 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939    9886  0   18  80   0 - 646405 futex_ 06:58 ?       00:00:01 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S 1000260+    8952    8939   28758  0   18  80   0 - 646405 futex_ 07:09 ?       00:00:00 /usr/bin/cluster-image-registry-operator --files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem --files=/etc/secrets/tls.crt --files=/etc/secrets/tls.key
1 S root        8998       1    8998  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e77869a3cd537d6f68ae21fefeac2b68942b8a972d0d38751c54c982b1b86813/userdata -c e77869a3cd537d6f68ae21fefeac2b68942b8a972d0d38751c54c982b1b86813 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-etcd-operator_etcd-operator-b45778765-qhsbc_10cb2fd4-636c-43cc-8c36-50cffb650f27/etcd-operator/0.log --log-level info -n k8s_etcd-operator_etcd-operator-b45778765-qhsbc_openshift-etcd-operator_10cb2fd4-636c-43cc-8c36-50cffb650f27_0 -P /run/containers/storage/overlay-containers/e77869a3cd537d6f68ae21fefeac2b68942b8a972d0d38751c54c982b1b86813/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e77869a3cd537d6f68ae21fefeac2b68942b8a972d0d38751c54c982b1b86813/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e77869a3cd537d6f68ae21fefeac2b68942b8a972d0d38751c54c982b1b86813/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e77869a3cd537d6f68ae21fefeac2b68942b8a972d0d38751c54c982b1b86813 -s
1 S root        9000       1    9000  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e5562c40eccf916179a067e707e83568cefc912272c915cba6d8acec49319221/userdata -c e5562c40eccf916179a067e707e83568cefc912272c915cba6d8acec49319221 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tk6p6_5a21c48a-596a-409d-8021-1425828a8a76/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_machine-api-operator-5694c8668f-tk6p6_openshift-machine-api_5a21c48a-596a-409d-8021-1425828a8a76_0 -P /run/containers/storage/overlay-containers/e5562c40eccf916179a067e707e83568cefc912272c915cba6d8acec49319221/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e5562c40eccf916179a067e707e83568cefc912272c915cba6d8acec49319221/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e5562c40eccf916179a067e707e83568cefc912272c915cba6d8acec49319221/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e5562c40eccf916179a067e707e83568cefc912272c915cba6d8acec49319221 -s
1 S root        9001       1    9001  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/cf1659d3992ed7da5b9df1a8b933c671d9b6d8d002125af9bf4276925fe84c5b/userdata -c cf1659d3992ed7da5b9df1a8b933c671d9b6d8d002125af9bf4276925fe84c5b --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-config-operator_openshift-config-operator-7777fb866f-vqfp6_3fdb5ac5-9260-4b3a-bbd6-30017ccfc191/openshift-config-operator/0.log --log-level info -n k8s_openshift-config-operator_openshift-config-operator-7777fb866f-vqfp6_openshift-config-operator_3fdb5ac5-9260-4b3a-bbd6-30017ccfc191_0 -P /run/containers/storage/overlay-containers/cf1659d3992ed7da5b9df1a8b933c671d9b6d8d002125af9bf4276925fe84c5b/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/cf1659d3992ed7da5b9df1a8b933c671d9b6d8d002125af9bf4276925fe84c5b/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/cf1659d3992ed7da5b9df1a8b933c671d9b6d8d002125af9bf4276925fe84c5b/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u cf1659d3992ed7da5b9df1a8b933c671d9b6d8d002125af9bf4276925fe84c5b -s
1 S root        9003       1    9003  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/bd64d59509c1a3720fbc07c20b42333ebfe35c06a9a581ec5c9338c144910e9a/userdata -c bd64d59509c1a3720fbc07c20b42333ebfe35c06a9a581ec5c9338c144910e9a --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-scheduler-operator_openshift-kube-scheduler-operator-5fdd9b5758-s5bl4_ba503e7e-7710-4be5-a871-ff39ff8b1296/kube-scheduler-operator-container/0.log --log-level info -n k8s_kube-scheduler-operator-container_openshift-kube-scheduler-operator-5fdd9b5758-s5bl4_openshift-kube-scheduler-operator_ba503e7e-7710-4be5-a871-ff39ff8b1296_0 -P /run/containers/storage/overlay-containers/bd64d59509c1a3720fbc07c20b42333ebfe35c06a9a581ec5c9338c144910e9a/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/bd64d59509c1a3720fbc07c20b42333ebfe35c06a9a581ec5c9338c144910e9a/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/bd64d59509c1a3720fbc07c20b42333ebfe35c06a9a581ec5c9338c144910e9a/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u bd64d59509c1a3720fbc07c20b42333ebfe35c06a9a581ec5c9338c144910e9a -s
1 S root        9007       1    9007  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/bd92a3916a440e5c655e1afac796466e653b50e47ba8f02ccad32f5de16f2d73/userdata -c bd92a3916a440e5c655e1afac796466e653b50e47ba8f02ccad32f5de16f2d73 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-console_downloads-7954f5f757-ng6vq_1c1c2c63-5588-4027-90a1-b78cdbe3b10b/download-server/0.log --log-level info -n k8s_download-server_downloads-7954f5f757-ng6vq_openshift-console_1c1c2c63-5588-4027-90a1-b78cdbe3b10b_0 -P /run/containers/storage/overlay-containers/bd92a3916a440e5c655e1afac796466e653b50e47ba8f02ccad32f5de16f2d73/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/bd92a3916a440e5c655e1afac796466e653b50e47ba8f02ccad32f5de16f2d73/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/bd92a3916a440e5c655e1afac796466e653b50e47ba8f02ccad32f5de16f2d73/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u bd92a3916a440e5c655e1afac796466e653b50e47ba8f02ccad32f5de16f2d73 -s
4 S nfsnobo+    9009    8998    9009  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:00 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9174  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:12 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9176  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:05 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9177  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:03 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9178  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:05 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9180  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:00 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9186  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:04 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9191  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:05 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9213  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:00 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
5 S nfsnobo+    9009    8998    9214  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:05 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9240  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:05 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
5 S nfsnobo+    9009    8998    9444  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:05 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9489  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:06 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9490  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:07 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9491  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:03 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9669  0   18  80   0 - 642948 ep_pol 06:58 ?       00:00:05 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998    9670  0   18  80   0 - 642948 futex_ 06:58 ?       00:00:05 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
1 S nfsnobo+    9009    8998   28301  0   18  80   0 - 642948 futex_ 07:09 ?       00:00:03 cluster-etcd-operator operator --config=/var/run/configmaps/config/config.yaml --terminate-on-files=/var/run/secrets/serving-cert/tls.crt --terminate-on-files=/var/run/secrets/serving-cert/tls.key --terminate-on-files=/var/run/secrets/etcd-client/tls.crt --terminate-on-files=/var/run/secrets/etcd-client/tls.key --terminate-on-files=/var/run/configmaps/etcd-ca/ca-bundle.crt --terminate-on-files=/var/run/configmaps/etcd-service-ca/service-ca.crt
4 S nfsnobo+    9011    9001    9011  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:00 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9256  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:08 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9259  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:00 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9260  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:05 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9261  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:00 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9270  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:00 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
5 S nfsnobo+    9011    9001    9383  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:05 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9538  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:01 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9539  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:06 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9540  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:07 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9799  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:04 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9848  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:00 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9849  0   15  80   0 - 583532 ep_pol 06:58 ?       00:00:02 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9850  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:03 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
1 S nfsnobo+    9011    9001    9851  0   15  80   0 - 583532 futex_ 06:58 ?       00:00:00 cluster-config-operator operator --operator-version=4.18.1 --authoritative-feature-gate-dir=/available-featuregates
4 S nfsnobo+    9024    9003    9024  0   17  80   0 - 622376 futex_ 06:58 ?       00:00:00 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003    9215  0   17  80   0 - 622376 futex_ 06:58 ?       00:00:10 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003    9216  0   17  80   0 - 622376 futex_ 06:58 ?       00:00:00 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003    9217  0   17  80   0 - 622376 futex_ 06:58 ?       00:00:02 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003    9219  0   17  80   0 - 622376 futex_ 06:58 ?       00:00:00 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
5 S nfsnobo+    9024    9003    9221  0   17  80   0 - 622376 ep_pol 06:58 ?       00:00:03 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003    9222  0   17  80   0 - 622376 futex_ 06:58 ?       00:00:01 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003    9223  0   17  80   0 - 622376 futex_ 06:58 ?       00:00:02 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003    9263  0   17  80   0 - 622376 futex_ 06:58 ?       00:00:03 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003    9328  0   17  80   0 - 622376 futex_ 06:58 ?       00:00:03 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003    9329  0   17  80   0 - 622376 futex_ 06:58 ?       00:00:02 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003    9527  0   17  80   0 - 622376 futex_ 06:58 ?       00:00:02 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003    9546  0   17  80   0 - 622376 futex_ 06:58 ?       00:00:03 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003   14217  0   17  80   0 - 622376 futex_ 06:59 ?       00:00:02 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003   14218  0   17  80   0 - 622376 futex_ 06:59 ?       00:00:03 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003   14219  0   17  80   0 - 622376 futex_ 06:59 ?       00:00:03 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9024    9003   14220  0   17  80   0 - 622376 futex_ 06:59 ?       00:00:02 cluster-kube-scheduler-operator operator --config=/var/run/configmaps/config/config.yaml
4 S 1000210+    9028    9000    9028  0   11  80   0 - 503862 ep_pol 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://localhost:8080/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S 1000210+    9028    9000    9229  0   11  80   0 - 503862 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://localhost:8080/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S 1000210+    9028    9000    9230  0   11  80   0 - 503862 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://localhost:8080/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S 1000210+    9028    9000    9231  0   11  80   0 - 503862 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://localhost:8080/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S 1000210+    9028    9000    9232  0   11  80   0 - 503862 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://localhost:8080/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S 1000210+    9028    9000    9233  0   11  80   0 - 503862 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://localhost:8080/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S 1000210+    9028    9000    9236  0   11  80   0 - 503862 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://localhost:8080/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S 1000210+    9028    9000    9237  0   11  80   0 - 503862 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://localhost:8080/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S 1000210+    9028    9000    9238  0   11  80   0 - 503862 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://localhost:8080/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S 1000210+    9028    9000    9241  0   11  80   0 - 503862 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://localhost:8080/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
1 S 1000210+    9028    9000  137328  0   11  80   0 - 503862 futex_ 08:21 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://localhost:8080/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --logtostderr=true --v=3
4 S 1000600+    9038    9007    9038  0  101  80   0 - 1767163 do_sel 06:58 ?      00:00:05 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11183  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11184  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11185  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11186  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11187  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11188  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11189  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11190  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11191  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11192  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11193  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11194  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11195  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11196  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11197  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11198  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11199  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11200  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11201  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11202  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11203  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11204  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11205  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11206  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11207  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11208  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11209  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11210  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11211  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11212  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11213  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11214  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11215  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11216  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11217  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11219  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11224  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11225  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11226  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11227  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11228  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11229  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11230  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11231  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11232  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11233  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11234  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11235  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11241  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11242  0  101  80   0 - 1767163 do_pol 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11243  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11244  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11245  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11246  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11247  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11248  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11251  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11252  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11253  0  101  80   0 - 1767163 do_pol 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11254  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11255  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11256  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11257  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11258  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11259  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11260  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11261  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11262  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11263  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11264  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11265  0  101  80   0 - 1767163 do_pol 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11266  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11267  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11268  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11269  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11270  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11275  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11276  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11277  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11278  0  101  80   0 - 1767163 do_pol 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11279  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11280  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11281  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11282  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11283  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11284  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11285  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11286  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11287  0  101  80   0 - 1767163 do_pol 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11288  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11289  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11290  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11291  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11292  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11293  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11294  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11295  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11296  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11297  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
5 S 1000600+    9038    9007   11298  0  101  80   0 - 1767163 inet_c 06:58 ?      00:00:00 python3 /tmp/serve.py
1 S root        9060       1    9060  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/175e332c2e4143ccb29a3f1ab121c76dbbd8933cf5e73035d4a68c3b1e6af547/userdata -c 175e332c2e4143ccb29a3f1ab121c76dbbd8933cf5e73035d4a68c3b1e6af547 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-machine-config-operator_machine-config-server-zdz9f_30fffcf6-c1f8-4fb4-9689-05fb42aaf2dd/machine-config-server/0.log --log-level info -n k8s_machine-config-server_machine-config-server-zdz9f_openshift-machine-config-operator_30fffcf6-c1f8-4fb4-9689-05fb42aaf2dd_0 -P /run/containers/storage/overlay-containers/175e332c2e4143ccb29a3f1ab121c76dbbd8933cf5e73035d4a68c3b1e6af547/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/175e332c2e4143ccb29a3f1ab121c76dbbd8933cf5e73035d4a68c3b1e6af547/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/175e332c2e4143ccb29a3f1ab121c76dbbd8933cf5e73035d4a68c3b1e6af547/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 175e332c2e4143ccb29a3f1ab121c76dbbd8933cf5e73035d4a68c3b1e6af547 -s
1 S root        9069       1    9069  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/035922dd826e0c02bf4eaedc785e26306e3df7507f90a36d5389ebd7e40bfb33/userdata -c 035922dd826e0c02bf4eaedc785e26306e3df7507f90a36d5389ebd7e40bfb33 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-zt5t2_d131a4aa-4055-4aa9-bfef-4354654e6577/cluster-samples-operator/0.log --log-level info -n k8s_cluster-samples-operator_cluster-samples-operator-665b6dd947-zt5t2_openshift-cluster-samples-operator_d131a4aa-4055-4aa9-bfef-4354654e6577_0 -P /run/containers/storage/overlay-containers/035922dd826e0c02bf4eaedc785e26306e3df7507f90a36d5389ebd7e40bfb33/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/035922dd826e0c02bf4eaedc785e26306e3df7507f90a36d5389ebd7e40bfb33/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/035922dd826e0c02bf4eaedc785e26306e3df7507f90a36d5389ebd7e40bfb33/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 035922dd826e0c02bf4eaedc785e26306e3df7507f90a36d5389ebd7e40bfb33 -s
4 S 1000280+    9072    9060    9072  0   15  80   0 - 580405 ep_pol 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9337  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9338  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9339  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9340  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9346  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9510  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9541  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9554  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9576  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9577  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9578  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9579  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9580  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
1 S 1000280+    9072    9060    9581  0   15  80   0 - 580405 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-server start --apiserver-url=https://api-int.crc.testing:6443 --payload-version=4.18.1 --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12
4 S 1000290+    9095    9069    9095  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
1 S 1000290+    9095    9069    9314  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:01 cluster-samples-operator
1 S 1000290+    9095    9069    9317  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
5 S 1000290+    9095    9069    9318  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:01 cluster-samples-operator
1 S 1000290+    9095    9069    9319  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
1 S 1000290+    9095    9069    9323  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
1 S 1000290+    9095    9069    9419  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
1 S 1000290+    9095    9069    9421  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
5 S 1000290+    9095    9069    9650  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
1 S 1000290+    9095    9069    9651  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
1 S 1000290+    9095    9069    9652  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:01 cluster-samples-operator
1 S 1000290+    9095    9069    9653  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
1 S 1000290+    9095    9069    9654  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
1 S 1000290+    9095    9069   11089  0   18  80   0 - 636870 ep_pol 06:58 ?       00:00:00 cluster-samples-operator
1 S 1000290+    9095    9069   11090  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
1 S 1000290+    9095    9069   11091  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
1 S 1000290+    9095    9069   11092  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
1 S 1000290+    9095    9069   11590  0   18  80   0 - 636870 futex_ 06:58 ?       00:00:00 cluster-samples-operator
1 S root        9300       1    9300  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/834dd4e22f29b5c2ea06c003eaac212884d0133b7f7d582c36920e9dd8c9de86/userdata -c 834dd4e22f29b5c2ea06c003eaac212884d0133b7f7d582c36920e9dd8c9de86 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-authentication-operator_authentication-operator-69f744f599-j9hjt_926a87b0-d967-45aa-8aff-13dcbf2c98e1/authentication-operator/0.log --log-level info -n k8s_authentication-operator_authentication-operator-69f744f599-j9hjt_openshift-authentication-operator_926a87b0-d967-45aa-8aff-13dcbf2c98e1_0 -P /run/containers/storage/overlay-containers/834dd4e22f29b5c2ea06c003eaac212884d0133b7f7d582c36920e9dd8c9de86/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/834dd4e22f29b5c2ea06c003eaac212884d0133b7f7d582c36920e9dd8c9de86/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/834dd4e22f29b5c2ea06c003eaac212884d0133b7f7d582c36920e9dd8c9de86/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 834dd4e22f29b5c2ea06c003eaac212884d0133b7f7d582c36920e9dd8c9de86 -s
1 S root        9301       1    9301  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/30a1b2d5ee4e0e3b22376eb56a9572a65ec62fb82a81dc58d957fd3e1ef62e7c/userdata -c 30a1b2d5ee4e0e3b22376eb56a9572a65ec62fb82a81dc58d957fd3e1ef62e7c --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-apiserver-operator_kube-apiserver-operator-766d6c64bb-hb9rt_498044ba-75d4-40c6-a7df-3846dc5b82d0/kube-apiserver-operator/0.log --log-level info -n k8s_kube-apiserver-operator_kube-apiserver-operator-766d6c64bb-hb9rt_openshift-kube-apiserver-operator_498044ba-75d4-40c6-a7df-3846dc5b82d0_0 -P /run/containers/storage/overlay-containers/30a1b2d5ee4e0e3b22376eb56a9572a65ec62fb82a81dc58d957fd3e1ef62e7c/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/30a1b2d5ee4e0e3b22376eb56a9572a65ec62fb82a81dc58d957fd3e1ef62e7c/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/30a1b2d5ee4e0e3b22376eb56a9572a65ec62fb82a81dc58d957fd3e1ef62e7c/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 30a1b2d5ee4e0e3b22376eb56a9572a65ec62fb82a81dc58d957fd3e1ef62e7c -s
1 S root        9302       1    9302  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/f8272b64a5ff7136080fa6dde0e7cf2bc5fa9c4052305e9cd259ace57f449394/userdata -c f8272b64a5ff7136080fa6dde0e7cf2bc5fa9c4052305e9cd259ace57f449394 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-cluster-machine-approver_machine-approver-56656f9798-f5ggg_db163501-b4f0-48b4-a558-e7f3d9d1c835/machine-approver-controller/0.log --log-level info -n k8s_machine-approver-controller_machine-approver-56656f9798-f5ggg_openshift-cluster-machine-approver_db163501-b4f0-48b4-a558-e7f3d9d1c835_0 -P /run/containers/storage/overlay-containers/f8272b64a5ff7136080fa6dde0e7cf2bc5fa9c4052305e9cd259ace57f449394/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/f8272b64a5ff7136080fa6dde0e7cf2bc5fa9c4052305e9cd259ace57f449394/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/f8272b64a5ff7136080fa6dde0e7cf2bc5fa9c4052305e9cd259ace57f449394/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u f8272b64a5ff7136080fa6dde0e7cf2bc5fa9c4052305e9cd259ace57f449394 -s
1 S root        9306       1    9306  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/6c86fb57c683fc04da8695a7cc10040d31651ac54d5c76ab6778431628d25c37/userdata -c 6c86fb57c683fc04da8695a7cc10040d31651ac54d5c76ab6778431628d25c37 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-9vcdh_dade3b21-b4f5-4558-bc6f-f63ef32cde34/multus-admission-controller/0.log --log-level info -n k8s_multus-admission-controller_multus-admission-controller-857f4d67dd-9vcdh_openshift-multus_dade3b21-b4f5-4558-bc6f-f63ef32cde34_0 -P /run/containers/storage/overlay-containers/6c86fb57c683fc04da8695a7cc10040d31651ac54d5c76ab6778431628d25c37/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/6c86fb57c683fc04da8695a7cc10040d31651ac54d5c76ab6778431628d25c37/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/6c86fb57c683fc04da8695a7cc10040d31651ac54d5c76ab6778431628d25c37/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 6c86fb57c683fc04da8695a7cc10040d31651ac54d5c76ab6778431628d25c37 -s
1 S root        9309       1    9309  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/40f088cbb8f4d7d3aa865e1aa547f8bcaff14900fca3632289b9944db9da14ca/userdata -c 40f088cbb8f4d7d3aa865e1aa547f8bcaff14900fca3632289b9944db9da14ca --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tk6p6_5a21c48a-596a-409d-8021-1425828a8a76/machine-api-operator/0.log --log-level info -n k8s_machine-api-operator_machine-api-operator-5694c8668f-tk6p6_openshift-machine-api_5a21c48a-596a-409d-8021-1425828a8a76_0 -P /run/containers/storage/overlay-containers/40f088cbb8f4d7d3aa865e1aa547f8bcaff14900fca3632289b9944db9da14ca/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/40f088cbb8f4d7d3aa865e1aa547f8bcaff14900fca3632289b9944db9da14ca/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/40f088cbb8f4d7d3aa865e1aa547f8bcaff14900fca3632289b9944db9da14ca/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 40f088cbb8f4d7d3aa865e1aa547f8bcaff14900fca3632289b9944db9da14ca -s
4 S root        9312    9300    9312  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:00 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300    9447  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:17 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300    9448  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:00 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
5 S root        9312    9300    9449  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:12 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
5 S root        9312    9300    9450  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:10 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300    9452  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:00 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300    9689  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:13 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300    9800  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:13 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300    9801  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:11 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300    9802  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:10 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300   10127  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:13 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300   10128  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:12 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300   10159  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:13 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300   10161  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:13 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300   10163  0   19  80   0 - 677129 ep_pol 06:58 ?       00:00:12 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300   11093  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:13 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300   11094  0   19  80   0 - 677129 futex_ 06:58 ?       00:00:12 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300   29144  0   19  80   0 - 677129 futex_ 07:09 ?       00:00:08 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
1 S root        9312    9300  173845  0   19  80   0 - 677129 futex_ 08:48 ?       00:00:01 authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate
4 S nfsnobo+    9313    9301    9313  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9401  0   18  80   0 - 676307 hrtime 06:58 ?       00:00:17 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9402  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:05 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
5 S nfsnobo+    9313    9301    9403  0   18  80   0 - 676307 ep_pol 06:58 ?       00:00:08 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9404  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:08 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9405  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9407  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:10 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9409  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9459  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:08 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9568  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:08 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9569  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:08 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9642  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:10 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9644  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:09 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9671  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:11 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
5 S nfsnobo+    9313    9301    9672  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:09 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9673  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:09 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9676  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:08 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9313    9301    9982  0   18  80   0 - 676307 futex_ 06:58 ?       00:00:06 cluster-kube-apiserver-operator operator --config=/var/run/configmaps/config/config.yaml
4 S nfsnobo+    9316    9306    9316  0   14  80   0 - 557043 futex_ 06:58 ?       00:00:01 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306    9423  0   14  80   0 - 557043 futex_ 06:58 ?       00:00:01 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306    9424  0   14  80   0 - 557043 futex_ 06:58 ?       00:00:00 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306    9425  0   14  80   0 - 557043 futex_ 06:58 ?       00:00:00 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306    9430  0   14  80   0 - 557043 futex_ 06:58 ?       00:00:00 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306    9432  0   14  80   0 - 557043 futex_ 06:58 ?       00:00:00 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306    9500  0   14  80   0 - 557043 futex_ 06:58 ?       00:00:00 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306    9565  0   14  80   0 - 557043 futex_ 06:58 ?       00:00:00 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306    9571  0   14  80   0 - 557043 ep_pol 06:58 ?       00:00:00 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306    9588  0   14  80   0 - 557043 futex_ 06:58 ?       00:00:00 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306   20716  0   14  80   0 - 557043 futex_ 07:04 ?       00:00:00 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306   40667  0   14  80   0 - 557043 futex_ 07:14 ?       00:00:00 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306   40668  0   14  80   0 - 557043 futex_ 07:14 ?       00:00:00 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
1 S nfsnobo+    9316    9306   65452  0   14  80   0 - 557043 futex_ 07:18 ?       00:00:00 /usr/bin/webhook -bind-address=0.0.0.0 -port=6443 -tls-private-key-file=/etc/webhook/tls.key -tls-cert-file=/etc/webhook/tls.crt -metrics-listen-address=127.0.0.1:9091 -alsologtostderr=true -ignore-namespaces=openshift-etcd,openshift-console,openshift-ingress-canary,openshift-apiserver,openshift-apiserver-operator,openshift-authentication,openshift-authentication-operator,openshift-cloud-network-config-controller,openshift-cluster-machine-approver,openshift-cluster-samples-operator,openshift-cluster-storage-operator,openshift-cluster-version,openshift-config-operator,openshift-console,openshift-console-operator,openshift-controller-manager,openshift-controller-manager-operator,openshift-dns,openshift-dns-operator,openshift-etcd-operator,openshift-image-registry,openshift-ingress,openshift-ingress-operator,openshift-kube-apiserver,openshift-kube-apiserver-operator,openshift-kube-controller-manager,openshift-kube-controller-manager-operator,openshift-kube-scheduler,openshift-kube-scheduler-operator,openshift-kube-storage-version-migrator,openshift-kube-storage-version-migrator-operator,openshift-machine-api,openshift-machine-config-operator,openshift-marketplace,openshift-monitoring,openshift-multus,openshift-network-console,openshift-network-diagnostics,openshift-network-node-identity,openshift-network-operator,openshift-oauth-apiserver,openshift-operator-lifecycle-manager,openshift-ovn-kubernetes,openshift-route-controller-manager,openshift-service-ca-operator,openshift-user-workload-monitoring
4 S root        9321    9302    9321  0   12  80   0 - 522989 futex_ 06:58 ?       00:00:00 /usr/bin/machine-approver --config=/var/run/configmaps/config/config.yaml -v=2 --logtostderr --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-cluster-machine-approver --api-group-version=machine.openshift.io/v1beta1 --max-concurrent-reconciles=10
1 S root        9321    9302    9473  0   12  80   0 - 522989 futex_ 06:58 ?       00:00:01 /usr/bin/machine-approver --config=/var/run/configmaps/config/config.yaml -v=2 --logtostderr --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-cluster-machine-approver --api-group-version=machine.openshift.io/v1beta1 --max-concurrent-reconciles=10
1 S root        9321    9302    9479  0   12  80   0 - 522989 futex_ 06:58 ?       00:00:00 /usr/bin/machine-approver --config=/var/run/configmaps/config/config.yaml -v=2 --logtostderr --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-cluster-machine-approver --api-group-version=machine.openshift.io/v1beta1 --max-concurrent-reconciles=10
1 S root        9321    9302    9480  0   12  80   0 - 522989 ep_pol 06:58 ?       00:00:00 /usr/bin/machine-approver --config=/var/run/configmaps/config/config.yaml -v=2 --logtostderr --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-cluster-machine-approver --api-group-version=machine.openshift.io/v1beta1 --max-concurrent-reconciles=10
1 S root        9321    9302    9481  0   12  80   0 - 522989 futex_ 06:58 ?       00:00:00 /usr/bin/machine-approver --config=/var/run/configmaps/config/config.yaml -v=2 --logtostderr --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-cluster-machine-approver --api-group-version=machine.openshift.io/v1beta1 --max-concurrent-reconciles=10
1 S root        9321    9302    9487  0   12  80   0 - 522989 futex_ 06:58 ?       00:00:00 /usr/bin/machine-approver --config=/var/run/configmaps/config/config.yaml -v=2 --logtostderr --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-cluster-machine-approver --api-group-version=machine.openshift.io/v1beta1 --max-concurrent-reconciles=10
1 S root        9321    9302    9517  0   12  80   0 - 522989 futex_ 06:58 ?       00:00:00 /usr/bin/machine-approver --config=/var/run/configmaps/config/config.yaml -v=2 --logtostderr --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-cluster-machine-approver --api-group-version=machine.openshift.io/v1beta1 --max-concurrent-reconciles=10
1 S root        9321    9302    9663  0   12  80   0 - 522989 futex_ 06:58 ?       00:00:00 /usr/bin/machine-approver --config=/var/run/configmaps/config/config.yaml -v=2 --logtostderr --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-cluster-machine-approver --api-group-version=machine.openshift.io/v1beta1 --max-concurrent-reconciles=10
1 S root        9321    9302    9664  0   12  80   0 - 522989 futex_ 06:58 ?       00:00:00 /usr/bin/machine-approver --config=/var/run/configmaps/config/config.yaml -v=2 --logtostderr --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-cluster-machine-approver --api-group-version=machine.openshift.io/v1beta1 --max-concurrent-reconciles=10
1 S root        9321    9302    9665  0   12  80   0 - 522989 futex_ 06:58 ?       00:00:00 /usr/bin/machine-approver --config=/var/run/configmaps/config/config.yaml -v=2 --logtostderr --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-cluster-machine-approver --api-group-version=machine.openshift.io/v1beta1 --max-concurrent-reconciles=10
1 S root        9321    9302    9784  0   12  80   0 - 522989 futex_ 06:58 ?       00:00:00 /usr/bin/machine-approver --config=/var/run/configmaps/config/config.yaml -v=2 --logtostderr --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-cluster-machine-approver --api-group-version=machine.openshift.io/v1beta1 --max-concurrent-reconciles=10
1 S root        9321    9302    9817  0   12  80   0 - 522989 futex_ 06:58 ?       00:00:00 /usr/bin/machine-approver --config=/var/run/configmaps/config/config.yaml -v=2 --logtostderr --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-cluster-machine-approver --api-group-version=machine.openshift.io/v1beta1 --max-concurrent-reconciles=10
4 S 1000210+    9324    9309    9324  0   11  80   0 - 1561641 futex_ 06:58 ?      00:00:00 /machine-api-operator start --images-json=/etc/machine-api-operator-config/images/images.json --alsologtostderr --v=3
1 S 1000210+    9324    9309    9512  0   11  80   0 - 1561641 futex_ 06:58 ?      00:00:02 /machine-api-operator start --images-json=/etc/machine-api-operator-config/images/images.json --alsologtostderr --v=3
1 S 1000210+    9324    9309    9513  0   11  80   0 - 1561641 futex_ 06:58 ?      00:00:00 /machine-api-operator start --images-json=/etc/machine-api-operator-config/images/images.json --alsologtostderr --v=3
1 S 1000210+    9324    9309    9514  0   11  80   0 - 1561641 futex_ 06:58 ?      00:00:00 /machine-api-operator start --images-json=/etc/machine-api-operator-config/images/images.json --alsologtostderr --v=3
1 S 1000210+    9324    9309    9516  0   11  80   0 - 1561641 futex_ 06:58 ?      00:00:00 /machine-api-operator start --images-json=/etc/machine-api-operator-config/images/images.json --alsologtostderr --v=3
1 S 1000210+    9324    9309    9518  0   11  80   0 - 1561641 futex_ 06:58 ?      00:00:00 /machine-api-operator start --images-json=/etc/machine-api-operator-config/images/images.json --alsologtostderr --v=3
1 S 1000210+    9324    9309    9668  0   11  80   0 - 1561641 ep_pol 06:58 ?      00:00:01 /machine-api-operator start --images-json=/etc/machine-api-operator-config/images/images.json --alsologtostderr --v=3
1 S 1000210+    9324    9309    9767  0   11  80   0 - 1561641 futex_ 06:58 ?      00:00:00 /machine-api-operator start --images-json=/etc/machine-api-operator-config/images/images.json --alsologtostderr --v=3
5 S 1000210+    9324    9309    9768  0   11  80   0 - 1561641 futex_ 06:58 ?      00:00:01 /machine-api-operator start --images-json=/etc/machine-api-operator-config/images/images.json --alsologtostderr --v=3
1 S 1000210+    9324    9309    9791  0   11  80   0 - 1561641 futex_ 06:58 ?      00:00:01 /machine-api-operator start --images-json=/etc/machine-api-operator-config/images/images.json --alsologtostderr --v=3
1 S 1000210+    9324    9309    9895  0   11  80   0 - 1561641 futex_ 06:58 ?      00:00:00 /machine-api-operator start --images-json=/etc/machine-api-operator-config/images/images.json --alsologtostderr --v=3
1 S root        9352       1    9352  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/a555b0d44a019748398b322739886d3b5904eb4e6ab43bd41df768b6d29464bf/userdata -c a555b0d44a019748398b322739886d3b5904eb4e6ab43bd41df768b6d29464bf --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-zt5t2_d131a4aa-4055-4aa9-bfef-4354654e6577/cluster-samples-operator-watch/0.log --log-level info -n k8s_cluster-samples-operator-watch_cluster-samples-operator-665b6dd947-zt5t2_openshift-cluster-samples-operator_d131a4aa-4055-4aa9-bfef-4354654e6577_0 -P /run/containers/storage/overlay-containers/a555b0d44a019748398b322739886d3b5904eb4e6ab43bd41df768b6d29464bf/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/a555b0d44a019748398b322739886d3b5904eb4e6ab43bd41df768b6d29464bf/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/a555b0d44a019748398b322739886d3b5904eb4e6ab43bd41df768b6d29464bf/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u a555b0d44a019748398b322739886d3b5904eb4e6ab43bd41df768b6d29464bf -s
4 S 1000290+    9357    9352    9357  0   11  80   0 - 507327 futex_ 06:58 ?       00:00:00 cluster-samples-operator-watch file-watcher-watchdog --namespace=openshift-cluster-samples-operator --process-name=cluster-samples-operator --termination-grace-period=30s --files=/etc/secrets/tls.crt,/etc/secrets/tls.key
1 S 1000290+    9357    9352    9472  0   11  80   0 - 507327 futex_ 06:58 ?       00:00:02 cluster-samples-operator-watch file-watcher-watchdog --namespace=openshift-cluster-samples-operator --process-name=cluster-samples-operator --termination-grace-period=30s --files=/etc/secrets/tls.crt,/etc/secrets/tls.key
1 S 1000290+    9357    9352    9474  0   11  80   0 - 507327 futex_ 06:58 ?       00:00:01 cluster-samples-operator-watch file-watcher-watchdog --namespace=openshift-cluster-samples-operator --process-name=cluster-samples-operator --termination-grace-period=30s --files=/etc/secrets/tls.crt,/etc/secrets/tls.key
1 S 1000290+    9357    9352    9475  0   11  80   0 - 507327 futex_ 06:58 ?       00:00:00 cluster-samples-operator-watch file-watcher-watchdog --namespace=openshift-cluster-samples-operator --process-name=cluster-samples-operator --termination-grace-period=30s --files=/etc/secrets/tls.crt,/etc/secrets/tls.key
1 S 1000290+    9357    9352    9476  0   11  80   0 - 507327 futex_ 06:58 ?       00:00:00 cluster-samples-operator-watch file-watcher-watchdog --namespace=openshift-cluster-samples-operator --process-name=cluster-samples-operator --termination-grace-period=30s --files=/etc/secrets/tls.crt,/etc/secrets/tls.key
1 S 1000290+    9357    9352    9477  0   11  80   0 - 507327 futex_ 06:58 ?       00:00:00 cluster-samples-operator-watch file-watcher-watchdog --namespace=openshift-cluster-samples-operator --process-name=cluster-samples-operator --termination-grace-period=30s --files=/etc/secrets/tls.crt,/etc/secrets/tls.key
5 S 1000290+    9357    9352    9496  0   11  80   0 - 507327 futex_ 06:58 ?       00:00:01 cluster-samples-operator-watch file-watcher-watchdog --namespace=openshift-cluster-samples-operator --process-name=cluster-samples-operator --termination-grace-period=30s --files=/etc/secrets/tls.crt,/etc/secrets/tls.key
1 S 1000290+    9357    9352    9497  0   11  80   0 - 507327 futex_ 06:58 ?       00:00:01 cluster-samples-operator-watch file-watcher-watchdog --namespace=openshift-cluster-samples-operator --process-name=cluster-samples-operator --termination-grace-period=30s --files=/etc/secrets/tls.crt,/etc/secrets/tls.key
1 S 1000290+    9357    9352    9498  0   11  80   0 - 507327 futex_ 06:58 ?       00:00:01 cluster-samples-operator-watch file-watcher-watchdog --namespace=openshift-cluster-samples-operator --process-name=cluster-samples-operator --termination-grace-period=30s --files=/etc/secrets/tls.crt,/etc/secrets/tls.key
1 S 1000290+    9357    9352    9504  0   11  80   0 - 507327 futex_ 06:58 ?       00:00:02 cluster-samples-operator-watch file-watcher-watchdog --namespace=openshift-cluster-samples-operator --process-name=cluster-samples-operator --termination-grace-period=30s --files=/etc/secrets/tls.crt,/etc/secrets/tls.key
1 S 1000290+    9357    9352   17284  0   11  80   0 - 507327 ep_pol 07:00 ?       00:00:00 cluster-samples-operator-watch file-watcher-watchdog --namespace=openshift-cluster-samples-operator --process-name=cluster-samples-operator --termination-grace-period=30s --files=/etc/secrets/tls.crt,/etc/secrets/tls.key
1 S root        9413       1    9413  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/4ffed818949dc564dc9bb968cefdd7ea5008e8c2a2e603dddb884f555a775f8c/userdata -c 4ffed818949dc564dc9bb968cefdd7ea5008e8c2a2e603dddb884f555a775f8c --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-operator-lifecycle-manager_packageserver-d55dfcdfc-hk6wz_b14e69b2-09a9-4af8-b903-6b3aeb219cd8/packageserver/0.log --log-level info -n k8s_packageserver_packageserver-d55dfcdfc-hk6wz_openshift-operator-lifecycle-manager_b14e69b2-09a9-4af8-b903-6b3aeb219cd8_0 -P /run/containers/storage/overlay-containers/4ffed818949dc564dc9bb968cefdd7ea5008e8c2a2e603dddb884f555a775f8c/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/4ffed818949dc564dc9bb968cefdd7ea5008e8c2a2e603dddb884f555a775f8c/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/4ffed818949dc564dc9bb968cefdd7ea5008e8c2a2e603dddb884f555a775f8c/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 4ffed818949dc564dc9bb968cefdd7ea5008e8c2a2e603dddb884f555a775f8c -s
4 S 1000330+    9418    9413    9418  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:00 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413    9525  0   18  80   0 - 687727 hrtime 06:58 ?       00:00:08 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413    9528  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:09 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413    9529  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:00 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413    9531  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:07 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413    9534  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:00 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413    9593  0   18  80   0 - 687727 ep_pol 06:58 ?       00:00:05 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
5 S 1000330+    9418    9413    9624  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:09 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413    9754  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:07 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413    9755  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:08 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413   10017  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:07 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413   10042  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:07 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413   10044  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:09 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413   10045  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:08 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413   10046  0   18  80   0 - 687727 futex_ 06:58 ?       00:00:09 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413   12521  0   18  80   0 - 687727 futex_ 06:59 ?       00:00:09 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413   12522  0   18  80   0 - 687727 futex_ 06:59 ?       00:00:09 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S 1000330+    9418    9413   12524  0   18  80   0 - 687727 futex_ 06:59 ?       00:00:08 /bin/package-server -v=4 --secure-port 5443 --global-namespace openshift-marketplace
1 S root        9507       1    9507  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/867a944ad599883018a76ce398fc92d9874e1626bbe22a393f4b92a42824eb12/userdata -c 867a944ad599883018a76ce398fc92d9874e1626bbe22a393f4b92a42824eb12 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-machine-config-operator_machine-config-controller-84d6567774-bwg2d_860c757e-a748-4689-aa5a-3414957b1d43/machine-config-controller/0.log --log-level info -n k8s_machine-config-controller_machine-config-controller-84d6567774-bwg2d_openshift-machine-config-operator_860c757e-a748-4689-aa5a-3414957b1d43_0 -P /run/containers/storage/overlay-containers/867a944ad599883018a76ce398fc92d9874e1626bbe22a393f4b92a42824eb12/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/867a944ad599883018a76ce398fc92d9874e1626bbe22a393f4b92a42824eb12/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/867a944ad599883018a76ce398fc92d9874e1626bbe22a393f4b92a42824eb12/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 867a944ad599883018a76ce398fc92d9874e1626bbe22a393f4b92a42824eb12 -s
4 S 1000280+    9509    9507    9509  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:02 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S 1000280+    9509    9507    9615  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:02 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
5 S 1000280+    9509    9507    9616  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:02 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
5 S 1000280+    9509    9507    9617  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:02 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S 1000280+    9509    9507    9618  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:01 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S 1000280+    9509    9507    9622  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:00 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S 1000280+    9509    9507    9731  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:01 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S 1000280+    9509    9507    9734  0   18  80   0 - 1711491 ep_pol 06:58 ?      00:00:02 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S 1000280+    9509    9507    9945  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:00 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S 1000280+    9509    9507    9946  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:00 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S 1000280+    9509    9507   10014  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:02 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
5 S 1000280+    9509    9507   10015  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:01 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
5 S 1000280+    9509    9507   10018  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:01 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
5 S 1000280+    9509    9507   10036  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:02 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S 1000280+    9509    9507   10153  0   18  80   0 - 1711491 futex_ 06:58 ?      00:00:00 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S 1000280+    9509    9507   14492  0   18  80   0 - 1711491 futex_ 06:59 ?      00:00:02 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S 1000280+    9509    9507   14493  0   18  80   0 - 1711491 futex_ 06:59 ?      00:00:01 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S 1000280+    9509    9507  111555  0   18  80   0 - 1711491 futex_ 07:57 ?      00:00:00 /usr/bin/machine-config-controller start --resourcelock-namespace=openshift-machine-config-operator --v=2 --payload-version=4.18.1
1 S root        9532       1    9532  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/b1b4de343b54fd97548c4ec76b8abee505f0e184cfebee5c3ea0a07d0bc95395/userdata -c b1b4de343b54fd97548c4ec76b8abee505f0e184cfebee5c3ea0a07d0bc95395 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-service-ca_service-ca-9c57cc56f-kkmch_5861cc2a-875b-459a-8940-9b16b85668a7/service-ca-controller/0.log --log-level info -n k8s_service-ca-controller_service-ca-9c57cc56f-kkmch_openshift-service-ca_5861cc2a-875b-459a-8940-9b16b85668a7_0 -P /run/containers/storage/overlay-containers/b1b4de343b54fd97548c4ec76b8abee505f0e184cfebee5c3ea0a07d0bc95395/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/b1b4de343b54fd97548c4ec76b8abee505f0e184cfebee5c3ea0a07d0bc95395/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/b1b4de343b54fd97548c4ec76b8abee505f0e184cfebee5c3ea0a07d0bc95395/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u b1b4de343b54fd97548c4ec76b8abee505f0e184cfebee5c3ea0a07d0bc95395 -s
4 S 1000500+    9536    9532    9536  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:00 service-ca-operator controller -v=2
1 S 1000500+    9536    9532    9594  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:05 service-ca-operator controller -v=2
1 S 1000500+    9536    9532    9595  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:00 service-ca-operator controller -v=2
1 S 1000500+    9536    9532    9596  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:00 service-ca-operator controller -v=2
5 S 1000500+    9536    9532    9598  0   17  80   0 - 653460 ep_pol 06:58 ?       00:00:02 service-ca-operator controller -v=2
1 S 1000500+    9536    9532    9599  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:01 service-ca-operator controller -v=2
1 S 1000500+    9536    9532    9603  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:00 service-ca-operator controller -v=2
5 S 1000500+    9536    9532    9783  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:03 service-ca-operator controller -v=2
5 S 1000500+    9536    9532    9980  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:04 service-ca-operator controller -v=2
5 S 1000500+    9536    9532   10064  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:02 service-ca-operator controller -v=2
1 S 1000500+    9536    9532   10065  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:01 service-ca-operator controller -v=2
1 S 1000500+    9536    9532   10143  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:01 service-ca-operator controller -v=2
1 S 1000500+    9536    9532   10144  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:01 service-ca-operator controller -v=2
1 S 1000500+    9536    9532   10145  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:00 service-ca-operator controller -v=2
1 S 1000500+    9536    9532   10146  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:01 service-ca-operator controller -v=2
1 S 1000500+    9536    9532   10147  0   17  80   0 - 653460 futex_ 06:58 ?       00:00:02 service-ca-operator controller -v=2
1 S 1000500+    9536    9532   29168  0   17  80   0 - 653460 futex_ 07:09 ?       00:00:02 service-ca-operator controller -v=2
1 S root        9559       1    9559  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/7be254d4b0e3286cba1883516ac8c1206148a869ab031d82c5b516343fc9fc14/userdata -c 7be254d4b0e3286cba1883516ac8c1206148a869ab031d82c5b516343fc9fc14 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jdh7s_d564c64d-4755-40aa-967d-8fb49704ef10/control-plane-machine-set-operator/0.log --log-level info -n k8s_control-plane-machine-set-operator_control-plane-machine-set-operator-78cbb6b69f-jdh7s_openshift-machine-api_d564c64d-4755-40aa-967d-8fb49704ef10_0 -P /run/containers/storage/overlay-containers/7be254d4b0e3286cba1883516ac8c1206148a869ab031d82c5b516343fc9fc14/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/7be254d4b0e3286cba1883516ac8c1206148a869ab031d82c5b516343fc9fc14/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/7be254d4b0e3286cba1883516ac8c1206148a869ab031d82c5b516343fc9fc14/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 7be254d4b0e3286cba1883516ac8c1206148a869ab031d82c5b516343fc9fc14 -s
4 S 1000210+    9564    9559    9564  0   12  80   0 - 523640 futex_ 06:58 ?       00:00:00 /manager -v=2 --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-machine-api
1 S 1000210+    9564    9559    9639  0   12  80   0 - 523640 hrtime 06:58 ?       00:00:02 /manager -v=2 --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-machine-api
1 S 1000210+    9564    9559    9640  0   12  80   0 - 523640 futex_ 06:58 ?       00:00:00 /manager -v=2 --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-machine-api
1 S 1000210+    9564    9559    9641  0   12  80   0 - 523640 futex_ 06:58 ?       00:00:01 /manager -v=2 --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-machine-api
1 S 1000210+    9564    9559    9643  0   12  80   0 - 523640 futex_ 06:58 ?       00:00:01 /manager -v=2 --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-machine-api
1 S 1000210+    9564    9559    9645  0   12  80   0 - 523640 futex_ 06:58 ?       00:00:00 /manager -v=2 --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-machine-api
1 S 1000210+    9564    9559    9835  0   12  80   0 - 523640 futex_ 06:58 ?       00:00:01 /manager -v=2 --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-machine-api
1 S 1000210+    9564    9559    9884  0   12  80   0 - 523640 futex_ 06:58 ?       00:00:00 /manager -v=2 --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-machine-api
1 S 1000210+    9564    9559    9885  0   12  80   0 - 523640 futex_ 06:58 ?       00:00:00 /manager -v=2 --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-machine-api
1 S 1000210+    9564    9559    9893  0   12  80   0 - 523640 futex_ 06:58 ?       00:00:00 /manager -v=2 --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-machine-api
1 S 1000210+    9564    9559    9894  0   12  80   0 - 523640 ep_pol 06:58 ?       00:00:00 /manager -v=2 --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-machine-api
1 S 1000210+    9564    9559    9939  0   12  80   0 - 523640 futex_ 06:58 ?       00:00:00 /manager -v=2 --leader-elect=true --leader-elect-lease-duration=137s --leader-elect-renew-deadline=107s --leader-elect-retry-period=26s --leader-elect-resource-namespace=openshift-machine-api
1 S root        9572       1    9572  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/697faf212502ad26c054c9367dc475df4db0e599c51240f6e2da3d18a790b6af/userdata -c 697faf212502ad26c054c9367dc475df4db0e599c51240f6e2da3d18a790b6af --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-machine-config-operator_machine-config-operator-74547568cd-7v4l7_a4a72d0a-d5a5-4c91-bee0-66af15c456e7/machine-config-operator/0.log --log-level info -n k8s_machine-config-operator_machine-config-operator-74547568cd-7v4l7_openshift-machine-config-operator_a4a72d0a-d5a5-4c91-bee0-66af15c456e7_0 -P /run/containers/storage/overlay-containers/697faf212502ad26c054c9367dc475df4db0e599c51240f6e2da3d18a790b6af/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/697faf212502ad26c054c9367dc475df4db0e599c51240f6e2da3d18a790b6af/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/697faf212502ad26c054c9367dc475df4db0e599c51240f6e2da3d18a790b6af/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 697faf212502ad26c054c9367dc475df4db0e599c51240f6e2da3d18a790b6af -s
4 S nfsnobo+    9575    9572    9575  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572    9656  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:05 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572    9657  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572    9658  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572    9659  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:02 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572    9660  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:01 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572    9661  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572    9786  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572    9997  0   15  80   0 - 582568 ep_pol 06:58 ?       00:00:00 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572   10000  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:02 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572   10001  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:02 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572   10010  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572   10011  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:02 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572   10031  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:00 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S nfsnobo+    9575    9572   10032  0   15  80   0 - 582568 futex_ 06:58 ?       00:00:02 /usr/bin/machine-config-operator start --images-json=/etc/mco/images/images.json --payload-version=4.18.1 --operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a
1 S root        9626       1    9626  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/bcb5f4d5b8b392d355d116982bfd89bdeb5dcdf49bc2d70d69f2ae7b3833799a/userdata -c bcb5f4d5b8b392d355d116982bfd89bdeb5dcdf49bc2d70d69f2ae7b3833799a --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-operator-lifecycle-manager_olm-operator-6b444d44fb-6gzbt_00918720-4227-4bee-932f-e4aa0d614be9/olm-operator/0.log --log-level info -n k8s_olm-operator_olm-operator-6b444d44fb-6gzbt_openshift-operator-lifecycle-manager_00918720-4227-4bee-932f-e4aa0d614be9_0 -P /run/containers/storage/overlay-containers/bcb5f4d5b8b392d355d116982bfd89bdeb5dcdf49bc2d70d69f2ae7b3833799a/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/bcb5f4d5b8b392d355d116982bfd89bdeb5dcdf49bc2d70d69f2ae7b3833799a/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/bcb5f4d5b8b392d355d116982bfd89bdeb5dcdf49bc2d70d69f2ae7b3833799a/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u bcb5f4d5b8b392d355d116982bfd89bdeb5dcdf49bc2d70d69f2ae7b3833799a -s
1 S root        9627       1    9627  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/a7051263e7c6b7eaba637af7b193d62b73e0c5b67db1115bf9c675ebf404c5a9/userdata -c a7051263e7c6b7eaba637af7b193d62b73e0c5b67db1115bf9c675ebf404c5a9 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-operator-lifecycle-manager_package-server-manager-789f6589d5-9hhnx_b785e264-9c98-4261-adbf-8f2182c2cbf7/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_package-server-manager-789f6589d5-9hhnx_openshift-operator-lifecycle-manager_b785e264-9c98-4261-adbf-8f2182c2cbf7_0 -P /run/containers/storage/overlay-containers/a7051263e7c6b7eaba637af7b193d62b73e0c5b67db1115bf9c675ebf404c5a9/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/a7051263e7c6b7eaba637af7b193d62b73e0c5b67db1115bf9c675ebf404c5a9/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/a7051263e7c6b7eaba637af7b193d62b73e0c5b67db1115bf9c675ebf404c5a9/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u a7051263e7c6b7eaba637af7b193d62b73e0c5b67db1115bf9c675ebf404c5a9 -s
1 S root        9630       1    9630  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/8e2692ce5df9b592c4c2eba177235380e42aa590bc09c5e7ac1c4a3f85cf5427/userdata -c 8e2692ce5df9b592c4c2eba177235380e42aa590bc09c5e7ac1c4a3f85cf5427 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-dns_dns-default-f9ncj_5891e675-9ac7-4773-b965-63556edc89b3/dns/0.log --log-level info -n k8s_dns_dns-default-f9ncj_openshift-dns_5891e675-9ac7-4773-b965-63556edc89b3_0 -P /run/containers/storage/overlay-containers/8e2692ce5df9b592c4c2eba177235380e42aa590bc09c5e7ac1c4a3f85cf5427/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/8e2692ce5df9b592c4c2eba177235380e42aa590bc09c5e7ac1c4a3f85cf5427/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/8e2692ce5df9b592c4c2eba177235380e42aa590bc09c5e7ac1c4a3f85cf5427/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 8e2692ce5df9b592c4c2eba177235380e42aa590bc09c5e7ac1c4a3f85cf5427 -s
4 S 1000330+    9632    9626    9632  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:00 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626    9683  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:39 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626    9684  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:28 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626    9685  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:25 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626    9687  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:24 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626    9690  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:00 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626    9811  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:27 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626    9871  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:28 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626    9897  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:24 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626    9898  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:00 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626    9899  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:25 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626   10076  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:24 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626   10080  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:25 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626   10089  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:27 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626   10090  0   18  80   0 - 672160 futex_ 06:58 ?       00:00:26 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626   26857  0   18  80   0 - 672160 futex_ 07:09 ?       00:00:25 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626   26858  0   18  80   0 - 672160 ep_pol 07:09 ?       00:00:22 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
1 S 1000330+    9632    9626   26996  0   18  80   0 - 672160 futex_ 07:09 ?       00:00:24 /bin/olm --namespace openshift-operator-lifecycle-manager --writeStatusName operator-lifecycle-manager --writePackageServerStatusName operator-lifecycle-manager-packageserver --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --protectedCopiedCSVNamespaces openshift
4 S root        9633    9630    9633  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:02 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630    9770  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:05 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630    9771  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:03 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630    9772  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:03 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630    9774  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:02 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630    9778  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:00 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630    9888  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:00 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630    9889  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:02 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630    9891  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:00 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630   10003  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:02 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630   10004  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:04 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630   10005  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:00 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630   10006  0   17  80   0 - 619003 futex_ 06:58 ?       00:00:01 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630   18098  0   17  80   0 - 619003 ep_pol 07:01 ?       00:00:02 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630   18099  0   17  80   0 - 619003 futex_ 07:01 ?       00:00:01 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630   18100  0   17  80   0 - 619003 futex_ 07:01 ?       00:00:01 coredns -conf /etc/coredns/Corefile
1 S root        9633    9630   33660  0   17  80   0 - 619003 futex_ 07:12 ?       00:00:02 coredns -conf /etc/coredns/Corefile
1 S root        9634       1    9634  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/18b8a694ddd841d3d33e388b0aba350be8acaa98ec42820e3aac4f264f76f410/userdata -c 18b8a694ddd841d3d33e388b0aba350be8acaa98ec42820e3aac4f264f76f410 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-operator-lifecycle-manager_catalog-operator-68c6474976-9jhmd_9c54b1b0-bdda-4d3f-87ed-6a4502756708/catalog-operator/0.log --log-level info -n k8s_catalog-operator_catalog-operator-68c6474976-9jhmd_openshift-operator-lifecycle-manager_9c54b1b0-bdda-4d3f-87ed-6a4502756708_0 -P /run/containers/storage/overlay-containers/18b8a694ddd841d3d33e388b0aba350be8acaa98ec42820e3aac4f264f76f410/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/18b8a694ddd841d3d33e388b0aba350be8acaa98ec42820e3aac4f264f76f410/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/18b8a694ddd841d3d33e388b0aba350be8acaa98ec42820e3aac4f264f76f410/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 18b8a694ddd841d3d33e388b0aba350be8acaa98ec42820e3aac4f264f76f410 -s
4 S 1000330+    9635    9627    9635  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:9090/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --logtostderr=true
1 S 1000330+    9635    9627    9709  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:9090/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --logtostderr=true
1 S 1000330+    9635    9627    9710  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:9090/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --logtostderr=true
1 S 1000330+    9635    9627    9711  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:9090/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --logtostderr=true
1 S 1000330+    9635    9627    9712  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:9090/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --logtostderr=true
1 S 1000330+    9635    9627    9713  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:9090/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --logtostderr=true
1 S 1000330+    9635    9627    9719  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:9090/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --logtostderr=true
1 S 1000330+    9635    9627    9781  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:9090/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --logtostderr=true
1 S 1000330+    9635    9627    9782  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:9090/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --logtostderr=true
1 S 1000330+    9635    9627   18756  0   11  80   0 - 503861 ep_pol 07:02 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:9090/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --logtostderr=true
1 S 1000330+    9635    9627   82794  0   11  80   0 - 503861 futex_ 07:29 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:9090/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key --logtostderr=true
4 S 1000330+    9637    9634    9637  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:00 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634    9773  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:04 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634    9775  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:05 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634    9776  0   18  80   0 - 671678 ep_pol 06:58 ?       00:00:03 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634    9777  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:03 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634    9779  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:00 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634    9890  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:00 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634    9896  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:04 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634   10022  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:02 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634   10023  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:01 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634   10091  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:04 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634   10092  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:03 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634   11095  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:03 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634   11096  0   18  80   0 - 671678 futex_ 06:58 ?       00:00:03 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634   13232  0   18  80   0 - 671678 futex_ 06:59 ?       00:00:04 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634   13233  0   18  80   0 - 671678 futex_ 06:59 ?       00:00:03 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634   13234  0   18  80   0 - 671678 futex_ 06:59 ?       00:00:02 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S 1000330+    9637    9634   26421  0   18  80   0 - 671678 futex_ 07:08 ?       00:00:03 /bin/catalog --namespace openshift-marketplace --configmapServerImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --opmImage=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad --util-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2 --writeStatusName operator-lifecycle-manager-catalog --tls-cert /srv-cert/tls.crt --tls-key /srv-cert/tls.key --client-ca /profile-collector-cert/tls.crt --set-workload-user-id=false
1 S root        9647       1    9647  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/d3914ad9d4866a1597af302b767b591765e10cd1c65aa571b512c373468bd39c/userdata -c d3914ad9d4866a1597af302b767b591765e10cd1c65aa571b512c373468bd39c --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-9vcdh_dade3b21-b4f5-4558-bc6f-f63ef32cde34/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_multus-admission-controller-857f4d67dd-9vcdh_openshift-multus_dade3b21-b4f5-4558-bc6f-f63ef32cde34_0 -P /run/containers/storage/overlay-containers/d3914ad9d4866a1597af302b767b591765e10cd1c65aa571b512c373468bd39c/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/d3914ad9d4866a1597af302b767b591765e10cd1c65aa571b512c373468bd39c/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/d3914ad9d4866a1597af302b767b591765e10cd1c65aa571b512c373468bd39c/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u d3914ad9d4866a1597af302b767b591765e10cd1c65aa571b512c373468bd39c -s
4 S nfsnobo+    9649    9647    9649  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/webhook/tls.key --tls-cert-file=/etc/webhook/tls.crt
1 S nfsnobo+    9649    9647    9699  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/webhook/tls.key --tls-cert-file=/etc/webhook/tls.crt
1 S nfsnobo+    9649    9647    9700  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/webhook/tls.key --tls-cert-file=/etc/webhook/tls.crt
1 S nfsnobo+    9649    9647    9701  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/webhook/tls.key --tls-cert-file=/etc/webhook/tls.crt
1 S nfsnobo+    9649    9647    9702  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/webhook/tls.key --tls-cert-file=/etc/webhook/tls.crt
1 S nfsnobo+    9649    9647    9703  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/webhook/tls.key --tls-cert-file=/etc/webhook/tls.crt
1 S nfsnobo+    9649    9647    9715  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/webhook/tls.key --tls-cert-file=/etc/webhook/tls.crt
1 S nfsnobo+    9649    9647    9749  0   11  80   0 - 503733 ep_pol 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/webhook/tls.key --tls-cert-file=/etc/webhook/tls.crt
1 S nfsnobo+    9649    9647    9751  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/webhook/tls.key --tls-cert-file=/etc/webhook/tls.crt
1 S nfsnobo+    9649    9647    9752  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/webhook/tls.key --tls-cert-file=/etc/webhook/tls.crt
1 S nfsnobo+    9649    9647   78668  0   11  80   0 - 503733 futex_ 07:25 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/webhook/tls.key --tls-cert-file=/etc/webhook/tls.crt
1 S root        9722       1    9722  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e54aa7f9f3efb8fc0b164bef53eb075f33011f05510b813887f7bb5c2d97f28d/userdata -c e54aa7f9f3efb8fc0b164bef53eb075f33011f05510b813887f7bb5c2d97f28d --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ingress-canary_ingress-canary-b8wfx_30567c0a-c27b-4c90-a327-2d0a427f92aa/serve-healthcheck-canary/0.log --log-level info -n k8s_serve-healthcheck-canary_ingress-canary-b8wfx_openshift-ingress-canary_30567c0a-c27b-4c90-a327-2d0a427f92aa_0 -P /run/containers/storage/overlay-containers/e54aa7f9f3efb8fc0b164bef53eb075f33011f05510b813887f7bb5c2d97f28d/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e54aa7f9f3efb8fc0b164bef53eb075f33011f05510b813887f7bb5c2d97f28d/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e54aa7f9f3efb8fc0b164bef53eb075f33011f05510b813887f7bb5c2d97f28d/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e54aa7f9f3efb8fc0b164bef53eb075f33011f05510b813887f7bb5c2d97f28d -s
1 S root        9724       1    9724  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/8c9685212ecbb287a4d6c756b2b063f615475f5cdf51f8484b41cd6bf9c895c1/userdata -c 8c9685212ecbb287a4d6c756b2b063f615475f5cdf51f8484b41cd6bf9c895c1 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-b67b599dd-6wn9g_938747a0-8051-4751-99a2-7b3167b23975/kube-storage-version-migrator-operator/0.log --log-level info -n k8s_kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-b67b599dd-6wn9g_openshift-kube-storage-version-migrator-operator_938747a0-8051-4751-99a2-7b3167b23975_0 -P /run/containers/storage/overlay-containers/8c9685212ecbb287a4d6c756b2b063f615475f5cdf51f8484b41cd6bf9c895c1/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/8c9685212ecbb287a4d6c756b2b063f615475f5cdf51f8484b41cd6bf9c895c1/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/8c9685212ecbb287a4d6c756b2b063f615475f5cdf51f8484b41cd6bf9c895c1/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 8c9685212ecbb287a4d6c756b2b063f615475f5cdf51f8484b41cd6bf9c895c1 -s
1 S root        9725       1    9725  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/2e952c74903802d45654080eaebd120eb0be5051709770dc938331909072ce91/userdata -c 2e952c74903802d45654080eaebd120eb0be5051709770dc938331909072ce91 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-controller-manager-operator_kube-controller-manager-operator-78b949d7b-fgh8h_a486ed64-5363-4a9a-95e1-d83d58920673/kube-controller-manager-operator/0.log --log-level info -n k8s_kube-controller-manager-operator_kube-controller-manager-operator-78b949d7b-fgh8h_openshift-kube-controller-manager-operator_a486ed64-5363-4a9a-95e1-d83d58920673_0 -P /run/containers/storage/overlay-containers/2e952c74903802d45654080eaebd120eb0be5051709770dc938331909072ce91/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/2e952c74903802d45654080eaebd120eb0be5051709770dc938331909072ce91/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/2e952c74903802d45654080eaebd120eb0be5051709770dc938331909072ce91/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 2e952c74903802d45654080eaebd120eb0be5051709770dc938331909072ce91 -s
1 S root        9726       1    9726  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/2f31cf52b2e043dca20ad0a33cfdb1dc2be97ace2dd0328ffee39ddcb7f009d1/userdata -c 2f31cf52b2e043dca20ad0a33cfdb1dc2be97ace2dd0328ffee39ddcb7f009d1 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-storage-version-migrator_migrator-59844c95c7-dw9sp_727e35fc-2fd0-4346-bff3-a086ef10c0c7/migrator/0.log --log-level info -n k8s_migrator_migrator-59844c95c7-dw9sp_openshift-kube-storage-version-migrator_727e35fc-2fd0-4346-bff3-a086ef10c0c7_0 -P /run/containers/storage/overlay-containers/2f31cf52b2e043dca20ad0a33cfdb1dc2be97ace2dd0328ffee39ddcb7f009d1/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/2f31cf52b2e043dca20ad0a33cfdb1dc2be97ace2dd0328ffee39ddcb7f009d1/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/2f31cf52b2e043dca20ad0a33cfdb1dc2be97ace2dd0328ffee39ddcb7f009d1/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 2f31cf52b2e043dca20ad0a33cfdb1dc2be97ace2dd0328ffee39ddcb7f009d1 -s
1 S root        9730       1    9730  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/f9e60b6c6055a7338a5f0bb3981b1820c262df1348e6b7899e7fef8f6e418aea/userdata -c f9e60b6c6055a7338a5f0bb3981b1820c262df1348e6b7899e7fef8f6e418aea --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-service-ca-operator_service-ca-operator-777779d784-898hd_46c30646-f384-458a-93b3-522895639fad/service-ca-operator/0.log --log-level info -n k8s_service-ca-operator_service-ca-operator-777779d784-898hd_openshift-service-ca-operator_46c30646-f384-458a-93b3-522895639fad_0 -P /run/containers/storage/overlay-containers/f9e60b6c6055a7338a5f0bb3981b1820c262df1348e6b7899e7fef8f6e418aea/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/f9e60b6c6055a7338a5f0bb3981b1820c262df1348e6b7899e7fef8f6e418aea/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/f9e60b6c6055a7338a5f0bb3981b1820c262df1348e6b7899e7fef8f6e418aea/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u f9e60b6c6055a7338a5f0bb3981b1820c262df1348e6b7899e7fef8f6e418aea -s
4 S nfsnobo+    9733    9725    9733  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:00 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725    9824  0   18  80   0 - 641432 hrtime 06:58 ?       00:00:13 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725    9825  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:03 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725    9826  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:06 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
5 S nfsnobo+    9733    9725    9827  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:04 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725    9831  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:00 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725    9838  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:06 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725    9845  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:00 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725    9846  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:05 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725    9992  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:03 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725    9993  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:03 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725    9994  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:03 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725    9995  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:03 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725   10149  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:03 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725   10150  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:04 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725   10156  0   18  80   0 - 641432 futex_ 06:58 ?       00:00:05 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725   14571  0   18  80   0 - 641432 futex_ 06:59 ?       00:00:03 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
1 S nfsnobo+    9733    9725   79845  0   18  80   0 - 641432 ep_pol 07:26 ?       00:00:03 cluster-kube-controller-manager-operator operator --config=/var/run/configmaps/config/config.yaml
4 S 1001        9736    9726    9736  0    1  80   0 -  1047 do_wai 06:58 ?        00:00:00 /bin/bash -c trap 'echo "Termination signal received, but ignored. Continuing..."; sleep infinity' TERM migrator "$@" & wait $! bash --alsologtostderr --v=2
4 S 1001        9743    9724    9743  0   17  80   0 - 621467 futex_ 06:58 ?       00:00:00 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724    9930  0   17  80   0 - 621467 futex_ 06:58 ?       00:00:04 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724    9931  0   17  80   0 - 621467 futex_ 06:58 ?       00:00:00 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724    9932  0   17  80   0 - 621467 futex_ 06:58 ?       00:00:03 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724    9933  0   17  80   0 - 621467 futex_ 06:58 ?       00:00:00 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724    9944  0   17  80   0 - 621467 futex_ 06:58 ?       00:00:00 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724   10034  0   17  80   0 - 621467 futex_ 06:58 ?       00:00:00 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
5 S 1001        9743    9724   10069  0   17  80   0 - 621467 ep_pol 06:58 ?       00:00:02 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724   10071  0   17  80   0 - 621467 futex_ 06:58 ?       00:00:00 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724   10106  0   17  80   0 - 621467 futex_ 06:58 ?       00:00:00 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724   10107  0   17  80   0 - 621467 futex_ 06:58 ?       00:00:00 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724   10108  0   17  80   0 - 621467 futex_ 06:58 ?       00:00:00 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724   10109  0   17  80   0 - 621467 futex_ 06:58 ?       00:00:03 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724   14476  0   17  80   0 - 621467 futex_ 06:59 ?       00:00:04 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724   14477  0   17  80   0 - 621467 futex_ 06:59 ?       00:00:00 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724   14478  0   17  80   0 - 621467 futex_ 06:59 ?       00:00:03 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
1 S 1001        9743    9724   14479  0   17  80   0 - 621467 futex_ 06:59 ?       00:00:03 cluster-kube-storage-version-migrator-operator start --config=/var/run/configmaps/config/config.yaml
4 S 1001        9745    9730    9745  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:00 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730    9876  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:06 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730    9877  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:02 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730    9878  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:00 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730    9879  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:00 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730    9880  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:00 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730    9892  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:00 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
5 S 1001        9745    9730    9947  0   15  80   0 - 582764 ep_pol 06:58 ?       00:00:02 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730    9948  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:02 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730    9976  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:03 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730   10047  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:02 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730   10083  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:01 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730   10100  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:00 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730   10102  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:00 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S 1001        9745    9730   10105  0   15  80   0 - 582764 futex_ 06:58 ?       00:00:03 service-ca-operator operator --config=/var/run/configmaps/config/operator-config.yaml -v=2
1 S root        9747       1    9747  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/1dad96ef60ef7ad2b6f5223a632ae430c2605a9c836516a934de074d7358c510/userdata -c 1dad96ef60ef7ad2b6f5223a632ae430c2605a9c836516a934de074d7358c510 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-machine-config-operator_machine-config-operator-74547568cd-7v4l7_a4a72d0a-d5a5-4c91-bee0-66af15c456e7/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_machine-config-operator-74547568cd-7v4l7_openshift-machine-config-operator_a4a72d0a-d5a5-4c91-bee0-66af15c456e7_0 -P /run/containers/storage/overlay-containers/1dad96ef60ef7ad2b6f5223a632ae430c2605a9c836516a934de074d7358c510/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/1dad96ef60ef7ad2b6f5223a632ae430c2605a9c836516a934de074d7358c510/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/1dad96ef60ef7ad2b6f5223a632ae430c2605a9c836516a934de074d7358c510/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 1dad96ef60ef7ad2b6f5223a632ae430c2605a9c836516a934de074d7358c510 -s
4 S nfsnobo+    9753    9747    9753  0   10  80   0 - 485428 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9753    9747    9806  0   10  80   0 - 485428 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9753    9747    9807  0   10  80   0 - 485428 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9753    9747    9808  0   10  80   0 - 485428 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9753    9747    9809  0   10  80   0 - 485428 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9753    9747    9810  0   10  80   0 - 485428 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9753    9747    9813  0   10  80   0 - 485428 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9753    9747    9814  0   10  80   0 - 485428 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9753    9747    9815  0   10  80   0 - 485428 ep_pol 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9753    9747    9855  0   10  80   0 - 485428 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S root        9756       1    9756  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/a91850045f25f0214e53cf609585b239ae411806f5feb1fd8ca89c8ccd950510/userdata -c a91850045f25f0214e53cf609585b239ae411806f5feb1fd8ca89c8ccd950510 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-machine-config-operator_machine-config-controller-84d6567774-bwg2d_860c757e-a748-4689-aa5a-3414957b1d43/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_machine-config-controller-84d6567774-bwg2d_openshift-machine-config-operator_860c757e-a748-4689-aa5a-3414957b1d43_0 -P /run/containers/storage/overlay-containers/a91850045f25f0214e53cf609585b239ae411806f5feb1fd8ca89c8ccd950510/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/a91850045f25f0214e53cf609585b239ae411806f5feb1fd8ca89c8ccd950510/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/a91850045f25f0214e53cf609585b239ae411806f5feb1fd8ca89c8ccd950510/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u a91850045f25f0214e53cf609585b239ae411806f5feb1fd8ca89c8ccd950510 -s
1 S root        9758       1    9758  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/66a9e00830c939175a7ff7168d2ead0ad93357713803817b0e2504624d3710b2/userdata -c 66a9e00830c939175a7ff7168d2ead0ad93357713803817b0e2504624d3710b2 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-oauth-apiserver_apiserver-7bbb656c7d-npnrn_3a82ff2b-e6a7-494f-a4f2-d98949f88eb8/oauth-apiserver/0.log --log-level info -n k8s_oauth-apiserver_apiserver-7bbb656c7d-npnrn_openshift-oauth-apiserver_3a82ff2b-e6a7-494f-a4f2-d98949f88eb8_0 -P /run/containers/storage/overlay-containers/66a9e00830c939175a7ff7168d2ead0ad93357713803817b0e2504624d3710b2/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/66a9e00830c939175a7ff7168d2ead0ad93357713803817b0e2504624d3710b2/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/66a9e00830c939175a7ff7168d2ead0ad93357713803817b0e2504624d3710b2/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 66a9e00830c939175a7ff7168d2ead0ad93357713803817b0e2504624d3710b2 -s
4 S 1000280+    9760    9756    9760  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000280+    9760    9756    9819  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000280+    9760    9756    9820  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000280+    9760    9756    9821  0   11  80   0 - 503861 ep_pol 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000280+    9760    9756    9822  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000280+    9760    9756    9823  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000280+    9760    9756    9836  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000280+    9760    9756    9837  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000280+    9760    9756    9843  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000280+    9760    9756    9853  0   11  80   0 - 503861 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S 1000280+    9760    9756   17265  0   11  80   0 - 503861 futex_ 07:00 ?       00:00:00 /usr/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
4 S 1000590+    9764    9722    9764  0    9  80   0 - 477176 futex_ 06:58 ?       00:00:00 ingress-operator serve-healthcheck
1 S 1000590+    9764    9722    9859  0    9  80   0 - 477176 futex_ 06:58 ?       00:00:00 ingress-operator serve-healthcheck
1 S 1000590+    9764    9722    9861  0    9  80   0 - 477176 futex_ 06:58 ?       00:00:00 ingress-operator serve-healthcheck
1 S 1000590+    9764    9722    9862  0    9  80   0 - 477176 futex_ 06:58 ?       00:00:00 ingress-operator serve-healthcheck
1 S 1000590+    9764    9722    9863  0    9  80   0 - 477176 futex_ 06:58 ?       00:00:00 ingress-operator serve-healthcheck
1 S 1000590+    9764    9722    9864  0    9  80   0 - 477176 futex_ 06:58 ?       00:00:00 ingress-operator serve-healthcheck
1 S 1000590+    9764    9722    9912  0    9  80   0 - 477176 futex_ 06:58 ?       00:00:00 ingress-operator serve-healthcheck
1 S 1000590+    9764    9722    9925  0    9  80   0 - 477176 ep_pol 06:58 ?       00:00:00 ingress-operator serve-healthcheck
1 S 1000590+    9764    9722    9940  0    9  80   0 - 477176 futex_ 06:58 ?       00:00:00 ingress-operator serve-healthcheck
4 S root        9766    9758    9766  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:00 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758    9904  0   17  80   0 - 638677 hrtime 06:58 ?       00:00:17 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
5 S root        9766    9758    9905  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:00 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758    9906  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:06 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758    9908  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:00 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758    9911  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:06 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758    9915  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:00 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758    9983  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:01 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
5 S root        9766    9758    9984  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:04 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758   10033  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:07 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758   10057  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:05 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758   10066  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:04 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758   10067  0   17  80   0 - 638677 ep_pol 06:58 ?       00:00:05 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758   10068  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:04 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758   10140  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:06 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758   10141  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:05 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9766    9758   10142  0   17  80   0 - 638677 futex_ 06:58 ?       00:00:00 oauth-apiserver start --secure-port=8443 --audit-log-path=/var/log/oauth-apiserver/audit.log --audit-log-format=json --audit-log-maxsize=100 --audit-log-maxbackup=10 --audit-policy-file=/var/run/configmaps/audit/policy.yaml --etcd-cafile=/var/run/configmaps/etcd-serving-ca/ca-bundle.crt --etcd-keyfile=/var/run/secrets/etcd-client/tls.key --etcd-certfile=/var/run/secrets/etcd-client/tls.crt --etcd-healthcheck-timeout=9s --etcd-readycheck-timeout=9s --shutdown-delay-duration=50s --shutdown-send-retry-after=true --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --enable-priority-and-fairness=false --api-audiences=https://kubernetes.default.svc --cors-allowed-origins=//127\.0\.0\.1(:|$) --cors-allowed-origins=//localhost(:|$) --etcd-servers=https://192.168.126.11:2379 --tls-cipher-suites=TLS_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --v=2
1 S root        9792       1    9792  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/dc039d1069e8e952e6c9cad8c0c86a3f09c329a60e6bd56c2fcdb844e9d2d568/userdata -c dc039d1069e8e952e6c9cad8c0c86a3f09c329a60e6bd56c2fcdb844e9d2d568 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-operator-lifecycle-manager_package-server-manager-789f6589d5-9hhnx_b785e264-9c98-4261-adbf-8f2182c2cbf7/package-server-manager/0.log --log-level info -n k8s_package-server-manager_package-server-manager-789f6589d5-9hhnx_openshift-operator-lifecycle-manager_b785e264-9c98-4261-adbf-8f2182c2cbf7_0 -P /run/containers/storage/overlay-containers/dc039d1069e8e952e6c9cad8c0c86a3f09c329a60e6bd56c2fcdb844e9d2d568/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/dc039d1069e8e952e6c9cad8c0c86a3f09c329a60e6bd56c2fcdb844e9d2d568/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/dc039d1069e8e952e6c9cad8c0c86a3f09c329a60e6bd56c2fcdb844e9d2d568/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u dc039d1069e8e952e6c9cad8c0c86a3f09c329a60e6bd56c2fcdb844e9d2d568 -s
4 S 1000330+    9798    9792    9798  0   19  80   0 - 650122 futex_ 06:58 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792    9903  0   19  80   0 - 650122 futex_ 06:58 ?       00:00:02 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792    9907  0   19  80   0 - 650122 futex_ 06:58 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
5 S 1000330+    9798    9792    9909  0   19  80   0 - 650122 futex_ 06:58 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792    9910  0   19  80   0 - 650122 futex_ 06:58 ?       00:00:00 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792    9920  0   19  80   0 - 650122 futex_ 06:58 ?       00:00:00 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792    9966  0   19  80   0 - 650122 futex_ 06:58 ?       00:00:00 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792    9978  0   19  80   0 - 650122 futex_ 06:58 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792   10174  0   19  80   0 - 650122 futex_ 06:58 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792   10195  0   19  80   0 - 650122 futex_ 06:58 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792   10355  0   19  80   0 - 650122 futex_ 06:58 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792   12050  0   19  80   0 - 650122 futex_ 06:59 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792   12051  0   19  80   0 - 650122 futex_ 06:59 ?       00:00:00 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792   12568  0   19  80   0 - 650122 futex_ 06:59 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792   12569  0   19  80   0 - 650122 futex_ 06:59 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792   12570  0   19  80   0 - 650122 ep_pol 06:59 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792   12571  0   19  80   0 - 650122 futex_ 06:59 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792   76959  0   19  80   0 - 650122 futex_ 07:23 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S 1000330+    9798    9792  101501  0   19  80   0 - 650122 futex_ 07:47 ?       00:00:01 /bin/psm start --name packageserver --namespace openshift-operator-lifecycle-manager --metrics=:9090
1 S root        9869       1    9869  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/76b0855d55a2f6adf8a238b073de9e7b05fdf7fe0b56b59f04deeb393556a2f0/userdata -c 76b0855d55a2f6adf8a238b073de9e7b05fdf7fe0b56b59f04deeb393556a2f0 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-dns-operator_dns-operator-744455d44c-c6zfm_ed125176-2266-4792-a978-f73f044b2d83/dns-operator/0.log --log-level info -n k8s_dns-operator_dns-operator-744455d44c-c6zfm_openshift-dns-operator_ed125176-2266-4792-a978-f73f044b2d83_0 -P /run/containers/storage/overlay-containers/76b0855d55a2f6adf8a238b073de9e7b05fdf7fe0b56b59f04deeb393556a2f0/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/76b0855d55a2f6adf8a238b073de9e7b05fdf7fe0b56b59f04deeb393556a2f0/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/76b0855d55a2f6adf8a238b073de9e7b05fdf7fe0b56b59f04deeb393556a2f0/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 76b0855d55a2f6adf8a238b073de9e7b05fdf7fe0b56b59f04deeb393556a2f0 -s
4 S nfsnobo+    9875    9869    9875  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:00 dns-operator
1 S nfsnobo+    9875    9869    9956  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:03 dns-operator
1 S nfsnobo+    9875    9869    9958  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:00 dns-operator
5 S nfsnobo+    9875    9869    9959  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:00 dns-operator
1 S nfsnobo+    9875    9869    9968  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:00 dns-operator
1 S nfsnobo+    9875    9869    9977  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:01 dns-operator
1 S nfsnobo+    9875    9869   10081  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:00 dns-operator
1 S nfsnobo+    9875    9869   10098  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:00 dns-operator
1 S nfsnobo+    9875    9869   10110  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:00 dns-operator
1 S nfsnobo+    9875    9869   10119  0   14  80   0 - 559972 ep_pol 06:58 ?       00:00:01 dns-operator
1 S nfsnobo+    9875    9869   10120  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:00 dns-operator
1 S nfsnobo+    9875    9869   10124  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:01 dns-operator
1 S nfsnobo+    9875    9869   10125  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:00 dns-operator
1 S nfsnobo+    9875    9869   10126  0   14  80   0 - 559972 futex_ 06:58 ?       00:00:00 dns-operator
4 S 1001        9883    9736    9883  0    9  80   0 - 462676 futex_ 06:58 ?       00:00:00 migrator --alsologtostderr --v=2
1 S 1001        9883    9736    9916  0    9  80   0 - 462676 futex_ 06:58 ?       00:00:00 migrator --alsologtostderr --v=2
1 S 1001        9883    9736    9918  0    9  80   0 - 462676 futex_ 06:58 ?       00:00:00 migrator --alsologtostderr --v=2
1 S 1001        9883    9736    9919  0    9  80   0 - 462676 futex_ 06:58 ?       00:00:00 migrator --alsologtostderr --v=2
5 S 1001        9883    9736    9921  0    9  80   0 - 462676 ep_pol 06:58 ?       00:00:00 migrator --alsologtostderr --v=2
1 S 1001        9883    9736    9923  0    9  80   0 - 462676 futex_ 06:58 ?       00:00:00 migrator --alsologtostderr --v=2
1 S 1001        9883    9736   10002  0    9  80   0 - 462676 futex_ 06:58 ?       00:00:00 migrator --alsologtostderr --v=2
5 S 1001        9883    9736   10009  0    9  80   0 - 462676 futex_ 06:58 ?       00:00:00 migrator --alsologtostderr --v=2
1 S 1001        9883    9736   29971  0    9  80   0 - 462676 futex_ 07:10 ?       00:00:00 migrator --alsologtostderr --v=2
1 S root        9913       1    9913  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/639edb224ecae4caa8d4cc5c08829f37908010e5b4bcc9f344ade72cc3daad34/userdata -c 639edb224ecae4caa8d4cc5c08829f37908010e5b4bcc9f344ade72cc3daad34 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-dns_dns-default-f9ncj_5891e675-9ac7-4773-b965-63556edc89b3/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_dns-default-f9ncj_openshift-dns_5891e675-9ac7-4773-b965-63556edc89b3_0 -P /run/containers/storage/overlay-containers/639edb224ecae4caa8d4cc5c08829f37908010e5b4bcc9f344ade72cc3daad34/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/639edb224ecae4caa8d4cc5c08829f37908010e5b4bcc9f344ade72cc3daad34/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/639edb224ecae4caa8d4cc5c08829f37908010e5b4bcc9f344ade72cc3daad34/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 639edb224ecae4caa8d4cc5c08829f37908010e5b4bcc9f344ade72cc3daad34 -s
4 S nfsnobo+    9924    9913    9924  0   11  80   0 - 503669 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9154 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:9153/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9924    9913    9962  0   11  80   0 - 503669 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9154 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:9153/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9924    9913    9963  0   11  80   0 - 503669 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9154 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:9153/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9924    9913    9964  0   11  80   0 - 503669 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9154 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:9153/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9924    9913    9965  0   11  80   0 - 503669 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9154 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:9153/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9924    9913    9967  0   11  80   0 - 503669 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9154 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:9153/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9924    9913    9974  0   11  80   0 - 503669 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9154 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:9153/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9924    9913    9975  0   11  80   0 - 503669 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9154 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:9153/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9924    9913    9987  0   11  80   0 - 503669 ep_pol 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9154 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:9153/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9924    9913   18933  0   11  80   0 - 503669 futex_ 07:02 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9154 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:9153/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9924    9913   20898  0   11  80   0 - 503669 futex_ 07:04 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9154 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:9153/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S root        9928       1    9928  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/345f7b1f555aadb58ab7bc0a601594c574b3712d1b981339d057ecad58411d80/userdata -c 345f7b1f555aadb58ab7bc0a601594c574b3712d1b981339d057ecad58411d80 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-kube-storage-version-migrator_migrator-59844c95c7-dw9sp_727e35fc-2fd0-4346-bff3-a086ef10c0c7/graceful-termination/0.log --log-level info -n k8s_graceful-termination_migrator-59844c95c7-dw9sp_openshift-kube-storage-version-migrator_727e35fc-2fd0-4346-bff3-a086ef10c0c7_0 -P /run/containers/storage/overlay-containers/345f7b1f555aadb58ab7bc0a601594c574b3712d1b981339d057ecad58411d80/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/345f7b1f555aadb58ab7bc0a601594c574b3712d1b981339d057ecad58411d80/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/345f7b1f555aadb58ab7bc0a601594c574b3712d1b981339d057ecad58411d80/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 345f7b1f555aadb58ab7bc0a601594c574b3712d1b981339d057ecad58411d80 -s
4 S 1001        9934    9928    9934  0    1  80   0 -  1176 do_wai 06:58 ?        00:00:00 /bin/bash -c trap 'echo "Gracefully sleeping for 25s to let another pod start..."; sleep 25; exit' EXIT while true; do echo "Waiting for termination..."; sleep 3600 & wait $!; done
1 S root        9985       1    9985  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/831cef1c92430f7916ccaece46217c5c81400ecf7a6fc6a0d3f01bd3506fbcd4/userdata -c 831cef1c92430f7916ccaece46217c5c81400ecf7a6fc6a0d3f01bd3506fbcd4 --exit-dir /var/run/crio/exits -l /var/log/pods/hostpath-provisioner_csi-hostpathplugin-q28kf_5c68d14b-8d32-49b2-b007-a65ede2c357c/hostpath-provisioner/0.log --log-level info -n k8s_hostpath-provisioner_csi-hostpathplugin-q28kf_hostpath-provisioner_5c68d14b-8d32-49b2-b007-a65ede2c357c_0 -P /run/containers/storage/overlay-containers/831cef1c92430f7916ccaece46217c5c81400ecf7a6fc6a0d3f01bd3506fbcd4/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/831cef1c92430f7916ccaece46217c5c81400ecf7a6fc6a0d3f01bd3506fbcd4/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/831cef1c92430f7916ccaece46217c5c81400ecf7a6fc6a0d3f01bd3506fbcd4/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 831cef1c92430f7916ccaece46217c5c81400ecf7a6fc6a0d3f01bd3506fbcd4 -s
4 S root        9990    9985    9990  0   17  80   0 - 477874 futex_ 06:58 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
1 S root        9990    9985   10084  0   17  80   0 - 477874 futex_ 06:58 ?       00:00:01 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
5 S root        9990    9985   10085  0   17  80   0 - 477874 futex_ 06:58 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
5 S root        9990    9985   10086  0   17  80   0 - 477874 futex_ 06:58 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
1 S root        9990    9985   10087  0   17  80   0 - 477874 futex_ 06:58 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
1 S root        9990    9985   10088  0   17  80   0 - 477874 futex_ 06:58 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
5 S root        9990    9985   10165  0   17  80   0 - 477874 futex_ 06:58 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
5 S root        9990    9985   10166  0   17  80   0 - 477874 futex_ 06:58 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
5 S root        9990    9985   10167  0   17  80   0 - 477874 futex_ 06:58 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
5 S root        9990    9985   12523  0   17  80   0 - 477874 ep_pol 06:59 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
5 S root        9990    9985   17204  0   17  80   0 - 477874 futex_ 07:00 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
5 S root        9990    9985   17205  0   17  80   0 - 477874 futex_ 07:00 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
5 S root        9990    9985   23510  0   17  80   0 - 477874 futex_ 07:07 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
1 S root        9990    9985   31949  0   17  80   0 - 477874 futex_ 07:11 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
5 S root        9990    9985   33621  0   17  80   0 - 477874 futex_ 07:12 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
1 S root        9990    9985   75460  0   17  80   0 - 477874 futex_ 07:22 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
1 S root        9990    9985   82079  0   17  80   0 - 477874 futex_ 07:28 ?       00:00:00 /usr/bin/hostpath-csi-driver --drivername=kubevirt.io.hostpath-provisioner --v=3 --datadir=[{"name":"local","path":"/csi-data-dir"}] --endpoint=unix:///csi/csi.sock --nodeid=crc --version=latest
1 S root        9991       1    9991  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/6ba6e5b3810fe87ee4edbe94a2dbb9157c0bc481c69dad9c2f49277451739597/userdata -c 6ba6e5b3810fe87ee4edbe94a2dbb9157c0bc481c69dad9c2f49277451739597 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-dns-operator_dns-operator-744455d44c-c6zfm_ed125176-2266-4792-a978-f73f044b2d83/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_dns-operator-744455d44c-c6zfm_openshift-dns-operator_ed125176-2266-4792-a978-f73f044b2d83_0 -P /run/containers/storage/overlay-containers/6ba6e5b3810fe87ee4edbe94a2dbb9157c0bc481c69dad9c2f49277451739597/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/6ba6e5b3810fe87ee4edbe94a2dbb9157c0bc481c69dad9c2f49277451739597/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/6ba6e5b3810fe87ee4edbe94a2dbb9157c0bc481c69dad9c2f49277451739597/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 6ba6e5b3810fe87ee4edbe94a2dbb9157c0bc481c69dad9c2f49277451739597 -s
4 S nfsnobo+    9998    9991    9998  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9998    9991   10024  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9998    9991   10025  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9998    9991   10026  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9998    9991   10028  0   11  80   0 - 503733 ep_pol 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9998    9991   10029  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9998    9991   10030  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9998    9991   10059  0   11  80   0 - 503733 futex_ 06:58 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9998    9991   17307  0   11  80   0 - 503733 futex_ 07:00 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9998    9991   17308  0   11  80   0 - 503733 futex_ 07:00 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S nfsnobo+    9998    9991  261692  0   11  80   0 - 503733 futex_ 09:13 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9393 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:60000/ --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key
1 S root       10060       1   10060  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/0ae6fa3729c84c938043565bd47ae5b0278b3929c3c13c8b2f2533c7a7281b5c/userdata -c 0ae6fa3729c84c938043565bd47ae5b0278b3929c3c13c8b2f2533c7a7281b5c --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-apiserver_apiserver-76f77b778f-zthm4_daa05454-65ea-4796-a2eb-79d127178570/openshift-apiserver/0.log --log-level info -n k8s_openshift-apiserver_apiserver-76f77b778f-zthm4_openshift-apiserver_daa05454-65ea-4796-a2eb-79d127178570_0 -P /run/containers/storage/overlay-containers/0ae6fa3729c84c938043565bd47ae5b0278b3929c3c13c8b2f2533c7a7281b5c/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/0ae6fa3729c84c938043565bd47ae5b0278b3929c3c13c8b2f2533c7a7281b5c/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/0ae6fa3729c84c938043565bd47ae5b0278b3929c3c13c8b2f2533c7a7281b5c/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 0ae6fa3729c84c938043565bd47ae5b0278b3929c3c13c8b2f2533c7a7281b5c -s
4 S root       10062   10060   10062  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:00 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10093  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:28 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
5 S root       10062   10060   10094  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:12 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10095  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:09 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10096  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:11 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10097  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:00 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10134  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:00 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
5 S root       10062   10060   10135  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:12 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10158  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:09 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10189  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:12 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10200  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:09 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10361  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:10 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10362  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:12 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10364  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:07 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10365  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:12 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
5 S root       10062   10060   10366  0   19  80   0 - 1772887 futex_ 06:58 ?      00:00:11 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   10495  0   19  80   0 - 1772887 ep_pol 06:58 ?      00:00:14 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060   26422  0   19  80   0 - 1772887 futex_ 07:08 ?      00:00:11 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10062   10060  218123  0   19  80   0 - 1772887 futex_ 09:02 ?      00:00:01 openshift-apiserver start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       10101       1   10101  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e395ef7a971d00cc109e315588859a6b660db409f51cfeaec6b5d6274ebaf977/userdata -c e395ef7a971d00cc109e315588859a6b660db409f51cfeaec6b5d6274ebaf977 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-apiserver_apiserver-76f77b778f-zthm4_daa05454-65ea-4796-a2eb-79d127178570/openshift-apiserver-check-endpoints/0.log --log-level info -n k8s_openshift-apiserver-check-endpoints_apiserver-76f77b778f-zthm4_openshift-apiserver_daa05454-65ea-4796-a2eb-79d127178570_0 -P /run/containers/storage/overlay-containers/e395ef7a971d00cc109e315588859a6b660db409f51cfeaec6b5d6274ebaf977/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e395ef7a971d00cc109e315588859a6b660db409f51cfeaec6b5d6274ebaf977/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e395ef7a971d00cc109e315588859a6b660db409f51cfeaec6b5d6274ebaf977/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e395ef7a971d00cc109e315588859a6b660db409f51cfeaec6b5d6274ebaf977 -s
4 S root       10104   10101   10104  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10114  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:04 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
5 S root       10104   10101   10115  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10116  0   18  80   0 - 658788 ep_pol 06:58 ?       00:00:01 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10117  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
5 S root       10104   10101   10121  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10131  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10132  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:01 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
5 S root       10104   10101   10133  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:01 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10170  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10171  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:03 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10531  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10532  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:03 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10533  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10534  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10535  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:02 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10578  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:00 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10104   10101   10579  0   18  80   0 - 658788 futex_ 06:58 ?       00:00:02 cluster-kube-apiserver-operator check-endpoints --listen 0.0.0.0:17698 --namespace openshift-apiserver --v 2
1 S root       10123       1   10123  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e9cbfc810e702f4bc9fc82da9d4f5dc7142f2b622bf474152e1cfb27610958fa/userdata -c e9cbfc810e702f4bc9fc82da9d4f5dc7142f2b622bf474152e1cfb27610958fa --exit-dir /var/run/crio/exits -l /var/log/pods/hostpath-provisioner_csi-hostpathplugin-q28kf_5c68d14b-8d32-49b2-b007-a65ede2c357c/node-driver-registrar/0.log --log-level info -n k8s_node-driver-registrar_csi-hostpathplugin-q28kf_hostpath-provisioner_5c68d14b-8d32-49b2-b007-a65ede2c357c_0 -P /run/containers/storage/overlay-containers/e9cbfc810e702f4bc9fc82da9d4f5dc7142f2b622bf474152e1cfb27610958fa/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e9cbfc810e702f4bc9fc82da9d4f5dc7142f2b622bf474152e1cfb27610958fa/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e9cbfc810e702f4bc9fc82da9d4f5dc7142f2b622bf474152e1cfb27610958fa/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e9cbfc810e702f4bc9fc82da9d4f5dc7142f2b622bf474152e1cfb27610958fa -s
4 S root       10130   10123   10130  0    8  80   0 - 314602 futex_ 06:58 ?       00:00:00 /usr/bin/csi-node-driver-registrar --v=3 --csi-address=/csi/csi.sock --kubelet-registration-path=/var/lib/kubelet/plugins/csi-hostpath/csi.sock
1 S root       10130   10123   10175  0    8  80   0 - 314602 futex_ 06:58 ?       00:00:00 /usr/bin/csi-node-driver-registrar --v=3 --csi-address=/csi/csi.sock --kubelet-registration-path=/var/lib/kubelet/plugins/csi-hostpath/csi.sock
1 S root       10130   10123   10176  0    8  80   0 - 314602 futex_ 06:58 ?       00:00:00 /usr/bin/csi-node-driver-registrar --v=3 --csi-address=/csi/csi.sock --kubelet-registration-path=/var/lib/kubelet/plugins/csi-hostpath/csi.sock
1 S root       10130   10123   10177  0    8  80   0 - 314602 futex_ 06:58 ?       00:00:00 /usr/bin/csi-node-driver-registrar --v=3 --csi-address=/csi/csi.sock --kubelet-registration-path=/var/lib/kubelet/plugins/csi-hostpath/csi.sock
1 S root       10130   10123   10181  0    8  80   0 - 314602 futex_ 06:58 ?       00:00:00 /usr/bin/csi-node-driver-registrar --v=3 --csi-address=/csi/csi.sock --kubelet-registration-path=/var/lib/kubelet/plugins/csi-hostpath/csi.sock
1 S root       10130   10123   10182  0    8  80   0 - 314602 futex_ 06:58 ?       00:00:00 /usr/bin/csi-node-driver-registrar --v=3 --csi-address=/csi/csi.sock --kubelet-registration-path=/var/lib/kubelet/plugins/csi-hostpath/csi.sock
1 S root       10130   10123   10194  0    8  80   0 - 314602 ep_pol 06:58 ?       00:00:00 /usr/bin/csi-node-driver-registrar --v=3 --csi-address=/csi/csi.sock --kubelet-registration-path=/var/lib/kubelet/plugins/csi-hostpath/csi.sock
1 S root       10130   10123   81898  0    8  80   0 - 314602 futex_ 07:28 ?       00:00:00 /usr/bin/csi-node-driver-registrar --v=3 --csi-address=/csi/csi.sock --kubelet-registration-path=/var/lib/kubelet/plugins/csi-hostpath/csi.sock
1 S root       10155       1   10155  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/5e7dc430863421212935ea4292ced772f3b551c7216391038c55a8791b883748/userdata -c 5e7dc430863421212935ea4292ced772f3b551c7216391038c55a8791b883748 --exit-dir /var/run/crio/exits -l /var/log/pods/hostpath-provisioner_csi-hostpathplugin-q28kf_5c68d14b-8d32-49b2-b007-a65ede2c357c/liveness-probe/0.log --log-level info -n k8s_liveness-probe_csi-hostpathplugin-q28kf_hostpath-provisioner_5c68d14b-8d32-49b2-b007-a65ede2c357c_0 -P /run/containers/storage/overlay-containers/5e7dc430863421212935ea4292ced772f3b551c7216391038c55a8791b883748/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/5e7dc430863421212935ea4292ced772f3b551c7216391038c55a8791b883748/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/5e7dc430863421212935ea4292ced772f3b551c7216391038c55a8791b883748/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 5e7dc430863421212935ea4292ced772f3b551c7216391038c55a8791b883748 -s
4 S root       10162   10155   10162  0   15  80   0 - 443560 ep_pol 06:58 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   10183  0   15  80   0 - 443560 futex_ 06:58 ?       00:00:01 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   10184  0   15  80   0 - 443560 futex_ 06:58 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   10185  0   15  80   0 - 443560 futex_ 06:58 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   10186  0   15  80   0 - 443560 futex_ 06:58 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   10197  0   15  80   0 - 443560 futex_ 06:58 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   10211  0   15  80   0 - 443560 futex_ 06:58 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   11127  0   15  80   0 - 443560 futex_ 06:58 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   11128  0   15  80   0 - 443560 futex_ 06:58 ?       00:00:01 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   11129  0   15  80   0 - 443560 futex_ 06:58 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   12436  0   15  80   0 - 443560 futex_ 06:59 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   18174  0   15  80   0 - 443560 futex_ 07:01 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   18175  0   15  80   0 - 443560 futex_ 07:01 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   18176  0   15  80   0 - 443560 futex_ 07:01 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10162   10155   77818  0   15  80   0 - 443560 futex_ 07:24 ?       00:00:00 /usr/bin/livenessprobe --csi-address=/csi/csi.sock --health-port=9898
1 S root       10190       1   10190  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/783bd49ba30bbe4796fb662ae84903517f4be1d4f587e06c7cc0b019be2e909a/userdata -c 783bd49ba30bbe4796fb662ae84903517f4be1d4f587e06c7cc0b019be2e909a --exit-dir /var/run/crio/exits -l /var/log/pods/hostpath-provisioner_csi-hostpathplugin-q28kf_5c68d14b-8d32-49b2-b007-a65ede2c357c/csi-provisioner/0.log --log-level info -n k8s_csi-provisioner_csi-hostpathplugin-q28kf_hostpath-provisioner_5c68d14b-8d32-49b2-b007-a65ede2c357c_0 -P /run/containers/storage/overlay-containers/783bd49ba30bbe4796fb662ae84903517f4be1d4f587e06c7cc0b019be2e909a/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/783bd49ba30bbe4796fb662ae84903517f4be1d4f587e06c7cc0b019be2e909a/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/783bd49ba30bbe4796fb662ae84903517f4be1d4f587e06c7cc0b019be2e909a/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 783bd49ba30bbe4796fb662ae84903517f4be1d4f587e06c7cc0b019be2e909a -s
4 S root       10192   10190   10192  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:01 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10309  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:02 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10312  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:00 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10313  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:00 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10315  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:00 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10460  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:00 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10517  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:00 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10576  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:00 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10580  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:00 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10593  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:00 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10594  0   14  80   0 - 433802 ep_pol 06:58 ?       00:00:00 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10595  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:00 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10596  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:00 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10192   10190   10597  0   14  80   0 - 433802 futex_ 06:58 ?       00:00:00 /usr/bin/csi-provisioner --v=5 --csi-address=/csi/csi.sock --feature-gates=Topology=true --enable-capacity=true --capacity-for-immediate-binding=true --extra-create-metadata=true --immediate-topology=false --strict-topology=true --node-deployment=true
1 S root       10477       1   10477  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/227899bec658339409b82752b5df8383501480225bee10639671abeb978c8725/userdata -c 227899bec658339409b82752b5df8383501480225bee10639671abeb978c8725 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/0.log --log-level info -n k8s_check-endpoints_network-check-source-55646444c4-trplf_openshift-network-diagnostics_9d751cbb-f2e2-430d-9754-c882a5e924a5_0 -P /run/containers/storage/overlay-containers/227899bec658339409b82752b5df8383501480225bee10639671abeb978c8725/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/227899bec658339409b82752b5df8383501480225bee10639671abeb978c8725/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/227899bec658339409b82752b5df8383501480225bee10639671abeb978c8725/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 227899bec658339409b82752b5df8383501480225bee10639671abeb978c8725 -s
4 S 1000460+   10479   10477   10479  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:00 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S 1000460+   10479   10477   10536  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:05 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S 1000460+   10479   10477   10537  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:02 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
5 S 1000460+   10479   10477   10538  0   17  80   0 - 636714 ep_pol 06:58 ?       00:00:02 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S 1000460+   10479   10477   10539  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:01 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S 1000460+   10479   10477   10540  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:00 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S 1000460+   10479   10477   10582  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:00 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
5 S 1000460+   10479   10477   10588  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:01 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
5 S 1000460+   10479   10477   10589  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:01 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
5 S 1000460+   10479   10477   10590  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:01 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S 1000460+   10479   10477   10694  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:00 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S 1000460+   10479   10477   10695  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:01 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S 1000460+   10479   10477   10728  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:01 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S 1000460+   10479   10477   10729  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:01 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S 1000460+   10479   10477   10730  0   17  80   0 - 636714 futex_ 06:58 ?       00:00:02 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S 1000460+   10479   10477   36690  0   17  80   0 - 636714 futex_ 07:13 ?       00:00:00 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S 1000460+   10479   10477   36691  0   17  80   0 - 636714 futex_ 07:13 ?       00:00:01 cluster-network-check-endpoints --listen 0.0.0.0:17698 --namespace openshift-network-diagnostics
1 S root       10501       1   10501  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/c15421d450926c611e195d2c3c056f46fc7766c87253dda03fe706e969d2a6bd/userdata -c c15421d450926c611e195d2c3c056f46fc7766c87253dda03fe706e969d2a6bd --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-network-console_networking-console-plugin-85b44fc459-gdk6g_5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/networking-console-plugin/0.log --log-level info -n k8s_networking-console-plugin_networking-console-plugin-85b44fc459-gdk6g_openshift-network-console_5fe485a1-e14f-4c09-b5b9-f252bc42b7e8_0 -P /run/containers/storage/overlay-containers/c15421d450926c611e195d2c3c056f46fc7766c87253dda03fe706e969d2a6bd/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/c15421d450926c611e195d2c3c056f46fc7766c87253dda03fe706e969d2a6bd/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/c15421d450926c611e195d2c3c056f46fc7766c87253dda03fe706e969d2a6bd/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u c15421d450926c611e195d2c3c056f46fc7766c87253dda03fe706e969d2a6bd -s
4 S 1000630+   10503   10501   10503  0    1  80   0 -  2628 sigsus 06:58 ?        00:00:00 nginx: master process nginx -c /tmp/nginx.conf -g daemon off;
1 S 1000630+   10528   10503   10528  0    1  80   0 -  2669 ep_pol 06:58 ?        00:00:00 nginx: worker process
1 S root       10545       1   10545  0    1  80   0 -  2078 do_pol 06:58 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/19af14ecb2bc49204b910fb09a36ddbfe97b49c8d0a9cc51e14e43662c41f415/userdata -c 19af14ecb2bc49204b910fb09a36ddbfe97b49c8d0a9cc51e14e43662c41f415 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-network-diagnostics_network-check-target-xd92c_3b6479f0-333b-4a96-9adf-2099afdc2447/network-check-target-container/0.log --log-level info -n k8s_network-check-target-container_network-check-target-xd92c_openshift-network-diagnostics_3b6479f0-333b-4a96-9adf-2099afdc2447_0 -P /run/containers/storage/overlay-containers/19af14ecb2bc49204b910fb09a36ddbfe97b49c8d0a9cc51e14e43662c41f415/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/19af14ecb2bc49204b910fb09a36ddbfe97b49c8d0a9cc51e14e43662c41f415/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/19af14ecb2bc49204b910fb09a36ddbfe97b49c8d0a9cc51e14e43662c41f415/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 19af14ecb2bc49204b910fb09a36ddbfe97b49c8d0a9cc51e14e43662c41f415 -s
4 S 1000460+   10553   10545   10553  0    8  80   0 - 437239 futex_ 06:58 ?       00:00:00 cluster-network-check-target
1 S 1000460+   10553   10545   10571  0    8  80   0 - 437239 futex_ 06:58 ?       00:00:00 cluster-network-check-target
1 S 1000460+   10553   10545   10572  0    8  80   0 - 437239 futex_ 06:58 ?       00:00:00 cluster-network-check-target
1 S 1000460+   10553   10545   10573  0    8  80   0 - 437239 futex_ 06:58 ?       00:00:00 cluster-network-check-target
1 S 1000460+   10553   10545   10574  0    8  80   0 - 437239 futex_ 06:58 ?       00:00:00 cluster-network-check-target
1 S 1000460+   10553   10545   10575  0    8  80   0 - 437239 futex_ 06:58 ?       00:00:00 cluster-network-check-target
1 S 1000460+   10553   10545  248629  0    8  80   0 - 437239 futex_ 09:09 ?       00:00:00 cluster-network-check-target
1 S 1000460+   10553   10545  254538  0    8  80   0 - 437239 ep_pol 09:11 ?       00:00:00 cluster-network-check-target
0 S core       11502       1   11502  0    1  80   0 -   271 signal 06:58 ?        00:00:00 catatonit -P
0 S core       11666    2078   11666  0    1  80   0 -  2633 ep_pol 06:58 ?        00:00:00 /usr/bin/dbus-broker-launch --scope user
0 S core       11668   11666   11668  0    1  80   0 -  1217 ep_pol 06:58 ?        00:00:01 dbus-broker --log 4 --controller 9 --machine-id 21801e6708c44f15b81395eb736a7cec --max-bytes 100000000000000 --max-fds 25000000000000 --max-matches 5000000000
1 S core       11837       1   11837  0    1  80   0 - 18600 ep_pol 06:58 ?        00:00:00 /usr/bin/pasta --config-net --dns-forward 169.254.0.1 -t none -u none -T none -U none --no-map-gw --quiet --netns /proc/11795/ns/net
1 S root       12185       1   12185  0    1  80   0 -  2078 do_pol 06:59 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e383522b56a11b5efe72b25a9e0b85797e43712c3219bb183a38829d6ab57d17/userdata -c e383522b56a11b5efe72b25a9e0b85797e43712c3219bb183a38829d6ab57d17 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-multus_network-metrics-daemon-drrxm_08c7f4c0-52b1-4047-9da9-c6a76b0e06e7/network-metrics-daemon/0.log --log-level info -n k8s_network-metrics-daemon_network-metrics-daemon-drrxm_openshift-multus_08c7f4c0-52b1-4047-9da9-c6a76b0e06e7_0 -P /run/containers/storage/overlay-containers/e383522b56a11b5efe72b25a9e0b85797e43712c3219bb183a38829d6ab57d17/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e383522b56a11b5efe72b25a9e0b85797e43712c3219bb183a38829d6ab57d17/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e383522b56a11b5efe72b25a9e0b85797e43712c3219bb183a38829d6ab57d17/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e383522b56a11b5efe72b25a9e0b85797e43712c3219bb183a38829d6ab57d17 -s
4 S nfsnobo+   12187   12185   12187  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12217  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12218  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12219  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:01 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12220  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12221  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12223  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12229  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12230  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12231  0   17  80   0 - 611670 ep_pol 06:59 ?       00:00:01 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12243  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12249  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12250  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12251  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12252  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12254  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S nfsnobo+   12187   12185   12255  0   17  80   0 - 611670 futex_ 06:59 ?       00:00:00 /usr/bin/network-metrics --node-name crc
1 S root       12225       1   12225  0    1  80   0 -  2078 do_pol 06:59 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/0eb54609b0502f95ace7043949810733328ed2d9d700de91d7500b5d1a07ae4c/userdata -c 0eb54609b0502f95ace7043949810733328ed2d9d700de91d7500b5d1a07ae4c --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-multus_network-metrics-daemon-drrxm_08c7f4c0-52b1-4047-9da9-c6a76b0e06e7/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_network-metrics-daemon-drrxm_openshift-multus_08c7f4c0-52b1-4047-9da9-c6a76b0e06e7_0 -P /run/containers/storage/overlay-containers/0eb54609b0502f95ace7043949810733328ed2d9d700de91d7500b5d1a07ae4c/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/0eb54609b0502f95ace7043949810733328ed2d9d700de91d7500b5d1a07ae4c/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/0eb54609b0502f95ace7043949810733328ed2d9d700de91d7500b5d1a07ae4c/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 0eb54609b0502f95ace7043949810733328ed2d9d700de91d7500b5d1a07ae4c -s
4 S nfsnobo+   12227   12225   12227  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12235  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12236  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12237  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12238  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12239  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12240  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12241  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12242  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12244  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12245  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12246  0   14  80   0 - 559224 ep_pol 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12247  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   12227   12225   12248  0   14  80   0 - 559224 futex_ 06:59 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:9091/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S root       15008       1   15008  0    1  80   0 -  2078 do_pol 06:59 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/0d3fd3db614915329e731713c5b9e6eee1d6874f5e981b3d6ea6f76d8e55cf1f/userdata -c 0d3fd3db614915329e731713c5b9e6eee1d6874f5e981b3d6ea6f76d8e55cf1f --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-image-registry_image-registry-66df7c8f76-p9mbq_fef7f840-17b5-4d91-ab11-54b31cdbee3a/registry/0.log --log-level info -n k8s_registry_image-registry-66df7c8f76-p9mbq_openshift-image-registry_fef7f840-17b5-4d91-ab11-54b31cdbee3a_0 -P /run/containers/storage/overlay-containers/0d3fd3db614915329e731713c5b9e6eee1d6874f5e981b3d6ea6f76d8e55cf1f/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/0d3fd3db614915329e731713c5b9e6eee1d6874f5e981b3d6ea6f76d8e55cf1f/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/0d3fd3db614915329e731713c5b9e6eee1d6874f5e981b3d6ea6f76d8e55cf1f/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 0d3fd3db614915329e731713c5b9e6eee1d6874f5e981b3d6ea6f76d8e55cf1f -s
4 S 1000260+   15010   15008   15010  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15029  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15030  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15031  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15032  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15033  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15034  0   15  80   0 - 581275 ep_pol 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15035  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15036  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15037  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15038  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:01 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15039  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15040  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15041  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S 1000260+   15010   15008   15042  0   15  80   0 - 581275 futex_ 06:59 ?       00:00:00 /usr/bin/dockerregistry
1 S root       15645       1   15645  0    1  80   0 -  2078 do_pol 06:59 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/f644e463d96e625278521c384a6af6e6eb9796247aa3e28d8f1cbac437b58f97/userdata -c f644e463d96e625278521c384a6af6e6eb9796247aa3e28d8f1cbac437b58f97 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-marketplace_marketplace-operator-79b997595-mxqpd_68a22413-3bbe-475d-a98e-09c99becb176/marketplace-operator/0.log --log-level info -n k8s_marketplace-operator_marketplace-operator-79b997595-mxqpd_openshift-marketplace_68a22413-3bbe-475d-a98e-09c99becb176_0 -P /run/containers/storage/overlay-containers/f644e463d96e625278521c384a6af6e6eb9796247aa3e28d8f1cbac437b58f97/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/f644e463d96e625278521c384a6af6e6eb9796247aa3e28d8f1cbac437b58f97/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/f644e463d96e625278521c384a6af6e6eb9796247aa3e28d8f1cbac437b58f97/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u f644e463d96e625278521c384a6af6e6eb9796247aa3e28d8f1cbac437b58f97 -s
4 S 1000170+   15647   15645   15647  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15675  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
5 S 1000170+   15647   15645   15676  0   19  80   0 - 650617 ep_pol 06:59 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15677  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:00 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15678  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:00 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15679  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15680  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15681  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15682  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15683  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15685  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:00 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15688  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:01 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15689  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15690  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15691  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:01 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15695  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   15698  0   19  80   0 - 650617 futex_ 06:59 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645   16710  0   19  80   0 - 650617 futex_ 07:00 ?       00:00:02 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S 1000170+   15647   15645  170812  0   19  80   0 - 650617 futex_ 08:46 ?       00:00:00 marketplace-operator -defaultsDir=/defaults -clusterOperatorName=marketplace -tls-cert /var/run/secrets/serving-cert/tls.crt -tls-key /var/run/secrets/serving-cert/tls.key
1 S root       16032       1   16032  0    1  80   0 -  2078 do_pol 06:59 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/7cc8eb856da3875594affb09b1522b4f97acaf389ce55ee223cc830a713e4c8f/userdata -c 7cc8eb856da3875594affb09b1522b4f97acaf389ce55ee223cc830a713e4c8f --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-marketplace_redhat-marketplace-9w6dm_c238f470-c89c-4992-8bb7-ad1c2cd58553/registry-server/0.log --log-level info -n k8s_registry-server_redhat-marketplace-9w6dm_openshift-marketplace_c238f470-c89c-4992-8bb7-ad1c2cd58553_0 -P /run/containers/storage/overlay-containers/7cc8eb856da3875594affb09b1522b4f97acaf389ce55ee223cc830a713e4c8f/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/7cc8eb856da3875594affb09b1522b4f97acaf389ce55ee223cc830a713e4c8f/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/7cc8eb856da3875594affb09b1522b4f97acaf389ce55ee223cc830a713e4c8f/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 7cc8eb856da3875594affb09b1522b4f97acaf389ce55ee223cc830a713e4c8f -s
4 S 1000170+   16035   16032   16035  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16069  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16070  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16071  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16072  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16073  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16076  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
5 S 1000170+   16035   16032   16077  0   19  80   0 - 1441808 ep_pol 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16078  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16079  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16080  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16081  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16082  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16083  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16084  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16085  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16086  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   16088  0   19  80   0 - 1441808 futex_ 06:59 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16035   16032   17310  0   19  80   0 - 1441808 futex_ 07:00 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S root       16398       1   16398  0    1  80   0 -  2078 do_pol 07:00 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/610ba3afb484a2f18488d7c37c79896e0013ec64487c54e52fcdea2eb1d94600/userdata -c 610ba3afb484a2f18488d7c37c79896e0013ec64487c54e52fcdea2eb1d94600 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-marketplace_redhat-operators-qfw9w_276bc9fb-9280-4ec8-8f8a-32f482040f97/registry-server/0.log --log-level info -n k8s_registry-server_redhat-operators-qfw9w_openshift-marketplace_276bc9fb-9280-4ec8-8f8a-32f482040f97_0 -P /run/containers/storage/overlay-containers/610ba3afb484a2f18488d7c37c79896e0013ec64487c54e52fcdea2eb1d94600/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/610ba3afb484a2f18488d7c37c79896e0013ec64487c54e52fcdea2eb1d94600/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/610ba3afb484a2f18488d7c37c79896e0013ec64487c54e52fcdea2eb1d94600/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 610ba3afb484a2f18488d7c37c79896e0013ec64487c54e52fcdea2eb1d94600 -s
4 S 1000170+   16402   16398   16402  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16432  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16433  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:05 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16434  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:05 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16437  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:05 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16438  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16439  0   19  80   0 - 1458401 ep_pol 07:00 ?      00:00:03 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
5 S 1000170+   16402   16398   16440  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:05 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16445  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16446  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16447  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:05 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16448  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:04 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16449  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:04 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16450  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:05 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16451  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:04 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16452  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:05 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16453  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:05 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   16631  0   19  80   0 - 1458401 futex_ 07:00 ?      00:00:04 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   16402   16398   82684  0   19  80   0 - 1458401 futex_ 07:29 ?      00:00:02 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S root       16959       1   16959  0    1  80   0 -  2078 do_pol 07:00 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/b7384599c1307c44f963a35a9a455190199c6e66027df5e370cd42b97d0da579/userdata -c b7384599c1307c44f963a35a9a455190199c6e66027df5e370cd42b97d0da579 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-authentication_oauth-openshift-846dc6fc5d-qh6cp_4dfa791b-d340-4d5c-a4f0-a4bc6001d98b/oauth-openshift/0.log --log-level info -n k8s_oauth-openshift_oauth-openshift-846dc6fc5d-qh6cp_openshift-authentication_4dfa791b-d340-4d5c-a4f0-a4bc6001d98b_0 -P /run/containers/storage/overlay-containers/b7384599c1307c44f963a35a9a455190199c6e66027df5e370cd42b97d0da579/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/b7384599c1307c44f963a35a9a455190199c6e66027df5e370cd42b97d0da579/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/b7384599c1307c44f963a35a9a455190199c6e66027df5e370cd42b97d0da579/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u b7384599c1307c44f963a35a9a455190199c6e66027df5e370cd42b97d0da579 -s
4 S root       16961   16959   16961  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:00 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16966  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:04 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16967  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:02 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
5 S root       16961   16959   16968  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:02 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16969  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:02 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16970  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:00 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16971  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:00 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16972  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:00 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16973  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:00 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16974  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:00 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16975  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:00 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16976  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:02 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16977  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:00 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16978  0   17  80   0 - 617697 ep_pol 07:00 ?       00:00:01 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16979  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:00 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   16980  0   17  80   0 - 617697 futex_ 07:00 ?       00:00:02 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
1 S root       16961   16959   18177  0   17  80   0 - 617697 futex_ 07:01 ?       00:00:02 oauth-server osinserver --config=/var/config/system/configmaps/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig --v=2 --audit-log-format=json --audit-log-maxbackup=10 --audit-log-maxsize=100 --audit-log-path=/var/log/oauth-server/audit.log --audit-policy-file=/var/run/configmaps/audit/audit.yaml
5 S dnsmasq    19975       1   19975  0    1  80   0 -  4778 do_pol 07:03 ?        00:00:02 /usr/sbin/dnsmasq
1 S root       24205       1   24205  0    1  80   0 -  2078 do_pol 07:08 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/a3c2816e09771ef396bd511dc6bbec94dc60ad0e4f902dc0527d78034153bd6f/userdata -c a3c2816e09771ef396bd511dc6bbec94dc60ad0e4f902dc0527d78034153bd6f --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nhk8j_004c338c-7f47-437c-945d-4c4f5fe66bc5/ovn-controller/0.log --log-level info -n k8s_ovn-controller_ovnkube-node-nhk8j_openshift-ovn-kubernetes_004c338c-7f47-437c-945d-4c4f5fe66bc5_0 -P /run/containers/storage/overlay-containers/a3c2816e09771ef396bd511dc6bbec94dc60ad0e4f902dc0527d78034153bd6f/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/a3c2816e09771ef396bd511dc6bbec94dc60ad0e4f902dc0527d78034153bd6f/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/a3c2816e09771ef396bd511dc6bbec94dc60ad0e4f902dc0527d78034153bd6f/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u a3c2816e09771ef396bd511dc6bbec94dc60ad0e4f902dc0527d78034153bd6f -s
4 S root       24208   24205   24208  0    5  80   0 - 83786 do_pol 07:08 ?        00:00:11 ovn-controller unix:/var/run/openvswitch/db.sock -vfile:off --no-chdir --pidfile=/var/run/ovn/ovn-controller.pid --syslog-method=null --log-file=/var/log/ovn/acl-audit-log.log -vFACILITY:local0 -vconsole:info -vconsole:acl_log:off -vPATTERN:console:%D{%Y-%m-%dT%H:%M:%S.###Z}|%05N|%c%T|%p|%m -vsyslog:acl_log:info -vfile:acl_log:info
1 S root       24208   24205   24218  0    5  80   0 - 83786 do_pol 07:08 ?        00:00:00 ovn-controller unix:/var/run/openvswitch/db.sock -vfile:off --no-chdir --pidfile=/var/run/ovn/ovn-controller.pid --syslog-method=null --log-file=/var/log/ovn/acl-audit-log.log -vFACILITY:local0 -vconsole:info -vconsole:acl_log:off -vPATTERN:console:%D{%Y-%m-%dT%H:%M:%S.###Z}|%05N|%c%T|%p|%m -vsyslog:acl_log:info -vfile:acl_log:info
1 S root       24208   24205   24219  0    5  80   0 - 83786 do_pol 07:08 ?        00:00:00 ovn-controller unix:/var/run/openvswitch/db.sock -vfile:off --no-chdir --pidfile=/var/run/ovn/ovn-controller.pid --syslog-method=null --log-file=/var/log/ovn/acl-audit-log.log -vFACILITY:local0 -vconsole:info -vconsole:acl_log:off -vPATTERN:console:%D{%Y-%m-%dT%H:%M:%S.###Z}|%05N|%c%T|%p|%m -vsyslog:acl_log:info -vfile:acl_log:info
1 S root       24208   24205   24220  0    5  80   0 - 83786 do_pol 07:08 ?        00:00:00 ovn-controller unix:/var/run/openvswitch/db.sock -vfile:off --no-chdir --pidfile=/var/run/ovn/ovn-controller.pid --syslog-method=null --log-file=/var/log/ovn/acl-audit-log.log -vFACILITY:local0 -vconsole:info -vconsole:acl_log:off -vPATTERN:console:%D{%Y-%m-%dT%H:%M:%S.###Z}|%05N|%c%T|%p|%m -vsyslog:acl_log:info -vfile:acl_log:info
1 S root       24208   24205   24221  0    5  80   0 - 83786 do_pol 07:08 ?        00:00:00 ovn-controller unix:/var/run/openvswitch/db.sock -vfile:off --no-chdir --pidfile=/var/run/ovn/ovn-controller.pid --syslog-method=null --log-file=/var/log/ovn/acl-audit-log.log -vFACILITY:local0 -vconsole:info -vconsole:acl_log:off -vPATTERN:console:%D{%Y-%m-%dT%H:%M:%S.###Z}|%05N|%c%T|%p|%m -vsyslog:acl_log:info -vfile:acl_log:info
1 S root       24226       1   24226  0    1  80   0 -  2078 do_pol 07:08 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/6320ea0b70119b4533fdee5082c72118221d82dd77b5ccd11121099ffc023ad6/userdata -c 6320ea0b70119b4533fdee5082c72118221d82dd77b5ccd11121099ffc023ad6 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nhk8j_004c338c-7f47-437c-945d-4c4f5fe66bc5/ovn-acl-logging/0.log --log-level info -n k8s_ovn-acl-logging_ovnkube-node-nhk8j_openshift-ovn-kubernetes_004c338c-7f47-437c-945d-4c4f5fe66bc5_0 -P /run/containers/storage/overlay-containers/6320ea0b70119b4533fdee5082c72118221d82dd77b5ccd11121099ffc023ad6/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/6320ea0b70119b4533fdee5082c72118221d82dd77b5ccd11121099ffc023ad6/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/6320ea0b70119b4533fdee5082c72118221d82dd77b5ccd11121099ffc023ad6/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 6320ea0b70119b4533fdee5082c72118221d82dd77b5ccd11121099ffc023ad6 -s
4 S root       24231   24226   24231  0    1  80   0 -  1113 do_wai 07:08 ?        00:00:00 /bin/bash -c set -euo pipefail . /ovnkube-lib/ovnkube-lib.sh || exit 1 start-audit-log-rotation 
4 S root       24268   24231   24268  0    1  80   0 -  1233 do_pol 07:08 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=tail /usr/bin/tail -F /var/log/ovn/acl-audit-log.log
1 S root       24278       1   24278  0    1  80   0 -  2078 do_pol 07:08 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/fde8db2fecf239600ecdbbc3b13a9e5b011b8fe2445acb4052b2ecef1975c221/userdata -c fde8db2fecf239600ecdbbc3b13a9e5b011b8fe2445acb4052b2ecef1975c221 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nhk8j_004c338c-7f47-437c-945d-4c4f5fe66bc5/kube-rbac-proxy-node/0.log --log-level info -n k8s_kube-rbac-proxy-node_ovnkube-node-nhk8j_openshift-ovn-kubernetes_004c338c-7f47-437c-945d-4c4f5fe66bc5_0 -P /run/containers/storage/overlay-containers/fde8db2fecf239600ecdbbc3b13a9e5b011b8fe2445acb4052b2ecef1975c221/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/fde8db2fecf239600ecdbbc3b13a9e5b011b8fe2445acb4052b2ecef1975c221/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/fde8db2fecf239600ecdbbc3b13a9e5b011b8fe2445acb4052b2ecef1975c221/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u fde8db2fecf239600ecdbbc3b13a9e5b011b8fe2445acb4052b2ecef1975c221 -s
4 S nfsnobo+   24280   24278   24280  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24310  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24311  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24312  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24313  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24319  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24320  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24321  0   15  80   0 - 577721 ep_pol 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24322  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24323  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24324  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24341  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24342  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24343  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24280   24278   24346  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9103 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29103/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S root       24334       1   24334  0    1  80   0 -  2078 do_pol 07:08 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/686e88e859a92e0d1de64b9346a693d581e09014dd6fd4c5e773be444b4c2200/userdata -c 686e88e859a92e0d1de64b9346a693d581e09014dd6fd4c5e773be444b4c2200 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nhk8j_004c338c-7f47-437c-945d-4c4f5fe66bc5/kube-rbac-proxy-ovn-metrics/0.log --log-level info -n k8s_kube-rbac-proxy-ovn-metrics_ovnkube-node-nhk8j_openshift-ovn-kubernetes_004c338c-7f47-437c-945d-4c4f5fe66bc5_0 -P /run/containers/storage/overlay-containers/686e88e859a92e0d1de64b9346a693d581e09014dd6fd4c5e773be444b4c2200/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/686e88e859a92e0d1de64b9346a693d581e09014dd6fd4c5e773be444b4c2200/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/686e88e859a92e0d1de64b9346a693d581e09014dd6fd4c5e773be444b4c2200/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 686e88e859a92e0d1de64b9346a693d581e09014dd6fd4c5e773be444b4c2200 -s
4 S nfsnobo+   24345   24334   24345  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24353  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24354  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24355  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24356  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24357  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24358  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24359  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24360  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24361  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24362  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24364  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24365  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24366  0   15  80   0 - 577721 ep_pol 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S nfsnobo+   24345   24334   24367  0   15  80   0 - 577721 futex_ 07:08 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9105 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 --upstream=http://127.0.0.1:29105/ --tls-private-key-file=/etc/pki/tls/metrics-cert/tls.key --tls-cert-file=/etc/pki/tls/metrics-cert/tls.crt
1 S root       24368       1   24368  0    1  80   0 -  2078 do_pol 07:08 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/13007047b201526afbe9b31abea4983e2717db2fd9f5523e476268e9138125af/userdata -c 13007047b201526afbe9b31abea4983e2717db2fd9f5523e476268e9138125af --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nhk8j_004c338c-7f47-437c-945d-4c4f5fe66bc5/northd/0.log --log-level info -n k8s_northd_ovnkube-node-nhk8j_openshift-ovn-kubernetes_004c338c-7f47-437c-945d-4c4f5fe66bc5_0 -P /run/containers/storage/overlay-containers/13007047b201526afbe9b31abea4983e2717db2fd9f5523e476268e9138125af/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/13007047b201526afbe9b31abea4983e2717db2fd9f5523e476268e9138125af/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/13007047b201526afbe9b31abea4983e2717db2fd9f5523e476268e9138125af/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 13007047b201526afbe9b31abea4983e2717db2fd9f5523e476268e9138125af -s
4 S root       24370   24368   24370  0    1  80   0 -  1113 do_wai 07:08 ?        00:00:00 /bin/bash -c set -xem if [[ -f /env/_master ]]; then   set -o allexport   source /env/_master   set +o allexport fi . /ovnkube-lib/ovnkube-lib.sh || exit 1  trap quit-ovn-northd TERM INT start-ovn-northd "${OVN_LOG_LEVEL}" 
4 S root       24383   24370   24383  0    3  80   0 - 44417 do_pol 07:08 ?        00:00:05 ovn-northd --no-chdir -vconsole:info -vfile:off -vPATTERN:console:%D{%Y-%m-%dT%H:%M:%S.###Z}|%05N|%c%T|%p|%m --pidfile /var/run/ovn/ovn-northd.pid --n-threads=1
1 S root       24383   24370   24384  0    3  80   0 - 44417 do_pol 07:08 ?        00:00:00 ovn-northd --no-chdir -vconsole:info -vfile:off -vPATTERN:console:%D{%Y-%m-%dT%H:%M:%S.###Z}|%05N|%c%T|%p|%m --pidfile /var/run/ovn/ovn-northd.pid --n-threads=1
1 S root       24383   24370   24385  0    3  80   0 - 44417 do_pol 07:08 ?        00:00:00 ovn-northd --no-chdir -vconsole:info -vfile:off -vPATTERN:console:%D{%Y-%m-%dT%H:%M:%S.###Z}|%05N|%c%T|%p|%m --pidfile /var/run/ovn/ovn-northd.pid --n-threads=1
1 S root       24420       1   24420  0    1  80   0 -  2078 do_pol 07:08 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/7ceea34c42eca00d295432da972ee39c85beb8929ef686a21fb43757d5e38f2e/userdata -c 7ceea34c42eca00d295432da972ee39c85beb8929ef686a21fb43757d5e38f2e --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nhk8j_004c338c-7f47-437c-945d-4c4f5fe66bc5/nbdb/0.log --log-level info -n k8s_nbdb_ovnkube-node-nhk8j_openshift-ovn-kubernetes_004c338c-7f47-437c-945d-4c4f5fe66bc5_0 -P /run/containers/storage/overlay-containers/7ceea34c42eca00d295432da972ee39c85beb8929ef686a21fb43757d5e38f2e/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/7ceea34c42eca00d295432da972ee39c85beb8929ef686a21fb43757d5e38f2e/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/7ceea34c42eca00d295432da972ee39c85beb8929ef686a21fb43757d5e38f2e/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 7ceea34c42eca00d295432da972ee39c85beb8929ef686a21fb43757d5e38f2e -s
4 S root       24422   24420   24422  0    1  80   0 -  1113 do_wai 07:08 ?        00:00:00 /bin/bash -c set -xem if [[ -f /env/_master ]]; then   set -o allexport   source /env/_master   set +o allexport fi . /ovnkube-lib/ovnkube-lib.sh || exit 1  trap quit-nbdb TERM INT start-nbdb ${OVN_LOG_LEVEL} 
4 S root       24443   24422   24443  0    1  80   0 -  5192 do_pol 07:08 ?        00:00:13 ovsdb-server -vconsole:info -vfile:off -vPATTERN:console:%D{%Y-%m-%dT%H:%M:%S.###Z}|%05N|%c%T|%p|%m --log-file=/var/log/ovn/ovsdb-server-nb.log --pidfile=/var/run/ovn/ovnnb_db.pid --remote=punix:/var/run/ovn/ovnnb_db.sock --unixctl=/var/run/ovn/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=db:OVN_Northbound,SSL,private_key --certificate=db:OVN_Northbound,SSL,certificate --ca-cert=db:OVN_Northbound,SSL,ca_cert --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers /etc/ovn/ovnnb_db.db
1 S root       24740       1   24740  0    1  80   0 -  2078 do_pol 07:08 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e961c44bc6a43b8480a56f69ad35875142cbdebd27bd64eeffab66fb24b7ee30/userdata -c e961c44bc6a43b8480a56f69ad35875142cbdebd27bd64eeffab66fb24b7ee30 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nhk8j_004c338c-7f47-437c-945d-4c4f5fe66bc5/sbdb/0.log --log-level info -n k8s_sbdb_ovnkube-node-nhk8j_openshift-ovn-kubernetes_004c338c-7f47-437c-945d-4c4f5fe66bc5_0 -P /run/containers/storage/overlay-containers/e961c44bc6a43b8480a56f69ad35875142cbdebd27bd64eeffab66fb24b7ee30/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e961c44bc6a43b8480a56f69ad35875142cbdebd27bd64eeffab66fb24b7ee30/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e961c44bc6a43b8480a56f69ad35875142cbdebd27bd64eeffab66fb24b7ee30/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e961c44bc6a43b8480a56f69ad35875142cbdebd27bd64eeffab66fb24b7ee30 -s
4 S root       24750   24740   24750  0    1  80   0 -  1113 do_wai 07:08 ?        00:00:00 /bin/bash -c set -xem if [[ -f /env/_master ]]; then   set -o allexport   source /env/_master   set +o allexport fi . /ovnkube-lib/ovnkube-lib.sh || exit 1  trap quit-sbdb TERM INT start-sbdb ${OVN_LOG_LEVEL} 
4 S root       24761   24750   24761  0    1  80   0 -  6957 do_pol 07:08 ?        00:00:12 ovsdb-server -vconsole:info -vfile:off -vPATTERN:console:%D{%Y-%m-%dT%H:%M:%S.###Z}|%05N|%c%T|%p|%m --log-file=/var/log/ovn/ovsdb-server-sb.log --pidfile=/var/run/ovn/ovnsb_db.pid --remote=punix:/var/run/ovn/ovnsb_db.sock --unixctl=/var/run/ovn/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=db:OVN_Southbound,SSL,private_key --certificate=db:OVN_Southbound,SSL,certificate --ca-cert=db:OVN_Southbound,SSL,ca_cert --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers /etc/ovn/ovnsb_db.db
1 S root       25091       1   25091  0    1  80   0 -  2078 do_pol 07:08 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/f8cbf2a0e732b8df1f5e089dddb05b686b7dc90aa240c7361981c08215ac9600/userdata -c f8cbf2a0e732b8df1f5e089dddb05b686b7dc90aa240c7361981c08215ac9600 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nhk8j_004c338c-7f47-437c-945d-4c4f5fe66bc5/ovnkube-controller/0.log --log-level info -n k8s_ovnkube-controller_ovnkube-node-nhk8j_openshift-ovn-kubernetes_004c338c-7f47-437c-945d-4c4f5fe66bc5_0 -P /run/containers/storage/overlay-containers/f8cbf2a0e732b8df1f5e089dddb05b686b7dc90aa240c7361981c08215ac9600/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/f8cbf2a0e732b8df1f5e089dddb05b686b7dc90aa240c7361981c08215ac9600/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/f8cbf2a0e732b8df1f5e089dddb05b686b7dc90aa240c7361981c08215ac9600/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u f8cbf2a0e732b8df1f5e089dddb05b686b7dc90aa240c7361981c08215ac9600 -s
4 S root       25093   25091   25093  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:00 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
1 S root       25093   25091   25131  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:18 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25132  0   30  80   0 - 926331 skb_wa 07:08 ?       00:00:11 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
1 S root       25093   25091   25133  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:00 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25134  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:07 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
1 S root       25093   25091   25135  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:00 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25136  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:02 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25137  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:06 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25138  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:00 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25139  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:00 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25140  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:05 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25141  0   30  80   0 - 926331 skb_wa 07:08 ?       00:00:04 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25142  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:01 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25143  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:02 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25144  0   30  80   0 - 926331 skb_wa 07:08 ?       00:00:07 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25145  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:02 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25146  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:09 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25222  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:05 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25410  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:06 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25411  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:00 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   25412  0   30  80   0 - 926331 futex_ 07:08 ?       00:00:00 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   26976  0   30  80   0 - 926331 futex_ 07:09 ?       00:00:08 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   32206  0   30  80   0 - 926331 futex_ 07:11 ?       00:00:07 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   34419  0   30  80   0 - 926331 futex_ 07:13 ?       00:00:08 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   34420  0   30  80   0 - 926331 futex_ 07:13 ?       00:00:09 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   34438  0   30  80   0 - 926331 ep_pol 07:13 ?       00:00:07 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   34439  0   30  80   0 - 926331 futex_ 07:13 ?       00:00:04 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   34646  0   30  80   0 - 926331 futex_ 07:13 ?       00:00:08 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   34649  0   30  80   0 - 926331 futex_ 07:13 ?       00:00:07 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
5 S root       25093   25091   34650  0   30  80   0 - 926331 skb_wa 07:13 ?       00:00:07 /usr/bin/ovnkube --init-ovnkube-controller crc --init-node crc --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode local --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-admin-network-policy --enable-multicast --zone crc --enable-interconnect --acl-logging-rate-limit 20 --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112
1 S root       25639       1   25639  0    1  80   0 -  2078 do_pol 07:08 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/909a37ace0afe1da0b5bcfee12098a2304a9e663ee9d7fb0757e49df990153fa/userdata -c 909a37ace0afe1da0b5bcfee12098a2304a9e663ee9d7fb0757e49df990153fa --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-multus_multus-8pnks_f9421086-70f1-441e-9aa0-5ac57a048c89/kube-multus/3.log --log-level info -n k8s_kube-multus_multus-8pnks_openshift-multus_f9421086-70f1-441e-9aa0-5ac57a048c89_3 -P /run/containers/storage/overlay-containers/909a37ace0afe1da0b5bcfee12098a2304a9e663ee9d7fb0757e49df990153fa/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/909a37ace0afe1da0b5bcfee12098a2304a9e663ee9d7fb0757e49df990153fa/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/909a37ace0afe1da0b5bcfee12098a2304a9e663ee9d7fb0757e49df990153fa/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 909a37ace0afe1da0b5bcfee12098a2304a9e663ee9d7fb0757e49df990153fa -s
4 S root       25641   25639   25641  0   24  80   0 - 742054 futex_ 07:08 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
1 S root       25641   25639   25660  0   24  80   0 - 742054 futex_ 07:08 ?       00:00:02 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   25661  0   24  80   0 - 742054 futex_ 07:08 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   25662  0   24  80   0 - 742054 futex_ 07:08 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   25663  0   24  80   0 - 742054 futex_ 07:08 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
1 S root       25641   25639   25664  0   24  80   0 - 742054 futex_ 07:08 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
1 S root       25641   25639   25665  0   24  80   0 - 742054 futex_ 07:08 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   25666  0   24  80   0 - 742054 futex_ 07:08 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   25667  0   24  80   0 - 742054 futex_ 07:08 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   25668  0   24  80   0 - 742054 futex_ 07:08 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   26852  0   24  80   0 - 742054 futex_ 07:09 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   26853  0   24  80   0 - 742054 futex_ 07:09 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   26854  0   24  80   0 - 742054 futex_ 07:09 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   26855  0   24  80   0 - 742054 futex_ 07:09 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   26856  0   24  80   0 - 742054 futex_ 07:09 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   34337  0   24  80   0 - 742054 futex_ 07:13 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   34339  0   24  80   0 - 742054 futex_ 07:13 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   34340  0   24  80   0 - 742054 futex_ 07:13 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   34352  0   24  80   0 - 742054 futex_ 07:13 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   34356  0   24  80   0 - 742054 futex_ 07:13 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   34532  0   24  80   0 - 742054 ep_pol 07:13 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   34562  0   24  80   0 - 742054 futex_ 07:13 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
5 S root       25641   25639   34566  0   24  80   0 - 742054 futex_ 07:13 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
1 S root       25641   25639   34673  0   24  80   0 - 742054 futex_ 07:13 ?       00:00:00 /usr/src/multus-cni/bin/multus-daemon
1 S root       26006       1   26006  0    1  80   0 -  2078 do_pol 07:08 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/5b220d2fd65b8935972ec4e009301b44d33fab066827ba944f10941239508de4/userdata -c 5b220d2fd65b8935972ec4e009301b44d33fab066827ba944f10941239508de4 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-route-controller-manager_route-controller-manager-648d56fc4f-9qltj_6944bc79-13b3-420b-ac7e-8642afef1647/route-controller-manager/0.log --log-level info -n k8s_route-controller-manager_route-controller-manager-648d56fc4f-9qltj_openshift-route-controller-manager_6944bc79-13b3-420b-ac7e-8642afef1647_0 -P /run/containers/storage/overlay-containers/5b220d2fd65b8935972ec4e009301b44d33fab066827ba944f10941239508de4/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/5b220d2fd65b8935972ec4e009301b44d33fab066827ba944f10941239508de4/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/5b220d2fd65b8935972ec4e009301b44d33fab066827ba944f10941239508de4/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 5b220d2fd65b8935972ec4e009301b44d33fab066827ba944f10941239508de4 -s
4 S 1000510+   26008   26006   26008  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:00 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26016  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:04 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26017  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:02 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26018  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:02 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26019  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:00 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26020  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:00 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26025  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:00 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26026  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:02 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26027  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:00 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
5 S 1000510+   26008   26006   26028  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:01 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26029  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:03 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26046  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:00 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26047  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:00 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26048  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:00 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26049  0   17  80   0 - 618904 ep_pol 07:08 ?       00:00:01 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26050  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:01 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000510+   26008   26006   26051  0   17  80   0 - 618904 futex_ 07:08 ?       00:00:03 route-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       26093       1   26093  0    1  80   0 -  2078 do_pol 07:08 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/286b869e705850ce1ea3595eac16e2d7dc3dd88985e314e83f81bf909d834d56/userdata -c 286b869e705850ce1ea3595eac16e2d7dc3dd88985e314e83f81bf909d834d56 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-controller-manager_controller-manager-b48b84df4-8p6gs_5b932d11-7bd7-493e-bb3b-3e3935c48a84/controller-manager/0.log --log-level info -n k8s_controller-manager_controller-manager-b48b84df4-8p6gs_openshift-controller-manager_5b932d11-7bd7-493e-bb3b-3e3935c48a84_0 -P /run/containers/storage/overlay-containers/286b869e705850ce1ea3595eac16e2d7dc3dd88985e314e83f81bf909d834d56/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/286b869e705850ce1ea3595eac16e2d7dc3dd88985e314e83f81bf909d834d56/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/286b869e705850ce1ea3595eac16e2d7dc3dd88985e314e83f81bf909d834d56/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 286b869e705850ce1ea3595eac16e2d7dc3dd88985e314e83f81bf909d834d56 -s
4 S 1000490+   26095   26093   26095  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:00 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26107  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:10 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26108  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:04 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26109  0   18  80   0 - 1715201 ep_pol 07:08 ?      00:00:02 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26110  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:04 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26111  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:00 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
5 S 1000490+   26095   26093   26112  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:02 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26113  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:04 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
5 S 1000490+   26095   26093   26114  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:04 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26115  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:06 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26116  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:03 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26117  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:00 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26118  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:02 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26119  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:03 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26120  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:04 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26121  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:02 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26123  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:05 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S 1000490+   26095   26093   26128  0   18  80   0 - 1715201 futex_ 07:08 ?      00:00:04 openshift-controller-manager start --config=/var/run/configmaps/config/config.yaml -v=2
1 S root       26977       1   26977  0    1  80   0 -  2078 do_pol 07:09 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/8e129d014f17a53415dbac1070d87aca4bbfd8ede60c95f4d56abb444ea22e30/userdata -c 8e129d014f17a53415dbac1070d87aca4bbfd8ede60c95f4d56abb444ea22e30 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-pcfqk_1a6bb43f-fea9-4fd4-86c6-50fe6845ec24/nmstate-operator/0.log --log-level info -n k8s_nmstate-operator_nmstate-operator-858ddd8f98-pcfqk_openshift-nmstate_1a6bb43f-fea9-4fd4-86c6-50fe6845ec24_0 -P /run/containers/storage/overlay-containers/8e129d014f17a53415dbac1070d87aca4bbfd8ede60c95f4d56abb444ea22e30/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/8e129d014f17a53415dbac1070d87aca4bbfd8ede60c95f4d56abb444ea22e30/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/8e129d014f17a53415dbac1070d87aca4bbfd8ede60c95f4d56abb444ea22e30/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 8e129d014f17a53415dbac1070d87aca4bbfd8ede60c95f4d56abb444ea22e30 -s
4 S 1000680+   26979   26977   26979  0   15  80   0 - 575064 ep_pol 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   26979   26977   26983  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:01 manager --zap-time-encoding=iso8601
1 S 1000680+   26979   26977   26984  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
5 S 1000680+   26979   26977   26985  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   26979   26977   26986  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   26979   26977   26987  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
5 S 1000680+   26979   26977   26988  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
5 S 1000680+   26979   26977   26989  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   26979   26977   26990  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   26979   26977   26991  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
5 S 1000680+   26979   26977   26992  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   26979   26977   26993  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   26979   26977   26994  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
5 S 1000680+   26979   26977   26995  0   15  80   0 - 575064 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   26979   26977   31611  0   15  80   0 - 575064 futex_ 07:11 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S root       27329       1   27329  0    1  80   0 -  2078 do_pol 07:09 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/20bc51df8697a8c5492dd194eaa8727c6b59457dbaca2f3aed434af32cbd4a2e/userdata -c 20bc51df8697a8c5492dd194eaa8727c6b59457dbaca2f3aed434af32cbd4a2e --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-console_console-bc65d5d98-4p6lt_7736fe75-c536-4ba3-a03c-d2df6d90f32b/console/0.log --log-level info -n k8s_console_console-bc65d5d98-4p6lt_openshift-console_7736fe75-c536-4ba3-a03c-d2df6d90f32b_0 -P /run/containers/storage/overlay-containers/20bc51df8697a8c5492dd194eaa8727c6b59457dbaca2f3aed434af32cbd4a2e/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/20bc51df8697a8c5492dd194eaa8727c6b59457dbaca2f3aed434af32cbd4a2e/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/20bc51df8697a8c5492dd194eaa8727c6b59457dbaca2f3aed434af32cbd4a2e/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 20bc51df8697a8c5492dd194eaa8727c6b59457dbaca2f3aed434af32cbd4a2e -s
4 S 1000600+   27331   27329   27331  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S 1000600+   27331   27329   27339  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:01 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
5 S 1000600+   27331   27329   27340  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:01 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S 1000600+   27331   27329   27341  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
5 S 1000600+   27331   27329   27342  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S 1000600+   27331   27329   27343  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S 1000600+   27331   27329   27344  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S 1000600+   27331   27329   27345  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S 1000600+   27331   27329   27346  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
5 S 1000600+   27331   27329   27347  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:01 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S 1000600+   27331   27329   27348  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S 1000600+   27331   27329   27349  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S 1000600+   27331   27329   27350  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S 1000600+   27331   27329   27351  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S 1000600+   27331   27329   27352  0   16  80   0 - 1656108 ep_pol 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S 1000600+   27331   27329   27353  0   16  80   0 - 1656108 futex_ 07:09 ?      00:00:00 /opt/bridge/bin/bridge --public-dir=/opt/bridge/static --config=/var/console-config/console-config.yaml --service-ca-file=/var/service-ca/service-ca.crt --v=2
1 S root       27465       1   27465  0    1  80   0 -  2078 do_pol 07:09 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/bdf7e158ab4e067f7920248ec013b16448dfc2d4eb17c081ebcbd4cf0c7aab5a/userdata -c bdf7e158ab4e067f7920248ec013b16448dfc2d4eb17c081ebcbd4cf0c7aab5a --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-9qx25_332c430e-f6e1-4068-8b12-d0a8bc3e1183/nmstate-webhook/0.log --log-level info -n k8s_nmstate-webhook_nmstate-webhook-6cdbc54649-9qx25_openshift-nmstate_332c430e-f6e1-4068-8b12-d0a8bc3e1183_0 -P /run/containers/storage/overlay-containers/bdf7e158ab4e067f7920248ec013b16448dfc2d4eb17c081ebcbd4cf0c7aab5a/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/bdf7e158ab4e067f7920248ec013b16448dfc2d4eb17c081ebcbd4cf0c7aab5a/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/bdf7e158ab4e067f7920248ec013b16448dfc2d4eb17c081ebcbd4cf0c7aab5a/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u bdf7e158ab4e067f7920248ec013b16448dfc2d4eb17c081ebcbd4cf0c7aab5a -s
1 S root       27467       1   27467  0    1  80   0 -  2078 do_pol 07:09 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/893ff85e4e13898d3bee761b02828541f110f49ef364f4e850516d5aefb0ab07/userdata -c 893ff85e4e13898d3bee761b02828541f110f49ef364f4e850516d5aefb0ab07 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-fc6mm_27a427fb-1f71-4935-89e7-764400c772c9/nmstate-metrics/0.log --log-level info -n k8s_nmstate-metrics_nmstate-metrics-fdff9cb8d-fc6mm_openshift-nmstate_27a427fb-1f71-4935-89e7-764400c772c9_0 -P /run/containers/storage/overlay-containers/893ff85e4e13898d3bee761b02828541f110f49ef364f4e850516d5aefb0ab07/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/893ff85e4e13898d3bee761b02828541f110f49ef364f4e850516d5aefb0ab07/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/893ff85e4e13898d3bee761b02828541f110f49ef364f4e850516d5aefb0ab07/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 893ff85e4e13898d3bee761b02828541f110f49ef364f4e850516d5aefb0ab07 -s
4 S 1000680+   27469   27465   27469  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   27485  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   27486  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   27487  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   27488  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   27489  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   27492  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   27493  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:01 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   27494  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   27495  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   27497  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   27498  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   28036  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   28037  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   28038  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   28142  0   17  80   0 - 611618 ep_pol 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27469   27465   28143  0   17  80   0 - 611618 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
4 S 1000680+   27470   27467   27470  0   11  80   0 - 500764 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27470   27467   27480  0   11  80   0 - 500764 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27470   27467   27481  0   11  80   0 - 500764 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27470   27467   27482  0   11  80   0 - 500764 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27470   27467   27483  0   11  80   0 - 500764 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27470   27467   27484  0   11  80   0 - 500764 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27470   27467   27490  0   11  80   0 - 500764 ep_pol 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27470   27467   27491  0   11  80   0 - 500764 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27470   27467   27501  0   11  80   0 - 500764 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27470   27467   27502  0   11  80   0 - 500764 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S 1000680+   27470   27467   31780  0   11  80   0 - 500764 futex_ 07:11 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S root       27499       1   27499  0    1  80   0 -  2078 do_pol 07:09 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/401af7e549fa0fbee8e15aebc53cd5a4d3eec0cc6a0525f4c719f30c9b0b9db4/userdata -c 401af7e549fa0fbee8e15aebc53cd5a4d3eec0cc6a0525f4c719f30c9b0b9db4 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-nmstate_nmstate-handler-8ztlc_98693660-7374-45d1-bc6c-677bb3532d3c/nmstate-handler/0.log --log-level info -n k8s_nmstate-handler_nmstate-handler-8ztlc_openshift-nmstate_98693660-7374-45d1-bc6c-677bb3532d3c_0 -P /run/containers/storage/overlay-containers/401af7e549fa0fbee8e15aebc53cd5a4d3eec0cc6a0525f4c719f30c9b0b9db4/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/401af7e549fa0fbee8e15aebc53cd5a4d3eec0cc6a0525f4c719f30c9b0b9db4/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/401af7e549fa0fbee8e15aebc53cd5a4d3eec0cc6a0525f4c719f30c9b0b9db4/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 401af7e549fa0fbee8e15aebc53cd5a4d3eec0cc6a0525f4c719f30c9b0b9db4 -s
4 S root       27503   27499   27503  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S root       27503   27499   27507  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:01 manager --zap-time-encoding=iso8601
5 S root       27503   27499   27508  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:01 manager --zap-time-encoding=iso8601
1 S root       27503   27499   27509  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:02 manager --zap-time-encoding=iso8601
5 S root       27503   27499   27510  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:01 manager --zap-time-encoding=iso8601
5 S root       27503   27499   27511  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:01 manager --zap-time-encoding=iso8601
1 S root       27503   27499   27512  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S root       27503   27499   27515  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:01 manager --zap-time-encoding=iso8601
5 S root       27503   27499   27516  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:02 manager --zap-time-encoding=iso8601
1 S root       27503   27499   27517  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:01 manager --zap-time-encoding=iso8601
1 S root       27503   27499   27518  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:01 manager --zap-time-encoding=iso8601
5 S root       27503   27499   27519  0   18  80   0 - 630307 ep_pol 07:09 ?       00:00:01 manager --zap-time-encoding=iso8601
1 S root       27503   27499   27520  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:00 manager --zap-time-encoding=iso8601
1 S root       27503   27499   27521  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:02 manager --zap-time-encoding=iso8601
5 S root       27503   27499   27522  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:01 manager --zap-time-encoding=iso8601
1 S root       27503   27499   27627  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:02 manager --zap-time-encoding=iso8601
1 S root       27503   27499   27628  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:01 manager --zap-time-encoding=iso8601
5 S root       27503   27499   28755  0   18  80   0 - 630307 futex_ 07:09 ?       00:00:02 manager --zap-time-encoding=iso8601
1 I root       27544       2   27544  0    1  60 -20 -     0 rescue 07:09 ?        00:00:00 [cfg80211]
1 S root       27549       1   27549  0    1  80   0 -  2078 do_pol 07:09 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/3eaf38feda99befb4a03168ed2cf9f4383f5757ef5ac22f129ade8d0c2b3fac7/userdata -c 3eaf38feda99befb4a03168ed2cf9f4383f5757ef5ac22f129ade8d0c2b3fac7 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-sjq8r_8d16767b-6c32-442f-b6b6-6c6cc78e2c25/nmstate-console-plugin/0.log --log-level info -n k8s_nmstate-console-plugin_nmstate-console-plugin-6b874cbd85-sjq8r_openshift-nmstate_8d16767b-6c32-442f-b6b6-6c6cc78e2c25_0 -P /run/containers/storage/overlay-containers/3eaf38feda99befb4a03168ed2cf9f4383f5757ef5ac22f129ade8d0c2b3fac7/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/3eaf38feda99befb4a03168ed2cf9f4383f5757ef5ac22f129ade8d0c2b3fac7/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/3eaf38feda99befb4a03168ed2cf9f4383f5757ef5ac22f129ade8d0c2b3fac7/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 3eaf38feda99befb4a03168ed2cf9f4383f5757ef5ac22f129ade8d0c2b3fac7 -s
4 S 1000680+   27552   27549   27552  0    1  80   0 -  2628 sigsus 07:09 ?        00:00:00 nginx: master process nginx -g daemon off;
1 S 1000680+   27557   27552   27557  0    1  80   0 -  2666 ep_pol 07:09 ?        00:00:00 nginx: worker process
1 S root       27631       1   27631  0    1  80   0 -  2078 do_pol 07:09 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/52d2b49be3a66723ba9899e0cf8a42232b377656370357d79c086645bebcb3bd/userdata -c 52d2b49be3a66723ba9899e0cf8a42232b377656370357d79c086645bebcb3bd --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-fc6mm_27a427fb-1f71-4935-89e7-764400c772c9/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_nmstate-metrics-fdff9cb8d-fc6mm_openshift-nmstate_27a427fb-1f71-4935-89e7-764400c772c9_0 -P /run/containers/storage/overlay-containers/52d2b49be3a66723ba9899e0cf8a42232b377656370357d79c086645bebcb3bd/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/52d2b49be3a66723ba9899e0cf8a42232b377656370357d79c086645bebcb3bd/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/52d2b49be3a66723ba9899e0cf8a42232b377656370357d79c086645bebcb3bd/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 52d2b49be3a66723ba9899e0cf8a42232b377656370357d79c086645bebcb3bd -s
4 S 1000680+   27633   27631   27633  0   15  80   0 - 577657 ep_pol 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27637  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27638  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27639  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27640  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27641  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27642  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27643  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27644  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27645  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27646  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27647  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27648  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27649  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S 1000680+   27633   27631   27650  0   15  80   0 - 577657 futex_ 07:09 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:8443 --upstream=http://127.0.0.1:8089
1 S root       29451       1   29451  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/2966519bd31da85fe6f42591426185fc21bd99d3e452d9b4f491530e2a0c03cb/userdata -c 2966519bd31da85fe6f42591426185fc21bd99d3e452d9b4f491530e2a0c03cb --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_metallb-operator-controller-manager-7d4655dff5-sw47j_bacb13e1-0d46-40d5-96a0-e765fdfd6a2c/manager/0.log --log-level info -n k8s_manager_metallb-operator-controller-manager-7d4655dff5-sw47j_metallb-system_bacb13e1-0d46-40d5-96a0-e765fdfd6a2c_0 -P /run/containers/storage/overlay-containers/2966519bd31da85fe6f42591426185fc21bd99d3e452d9b4f491530e2a0c03cb/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/2966519bd31da85fe6f42591426185fc21bd99d3e452d9b4f491530e2a0c03cb/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/2966519bd31da85fe6f42591426185fc21bd99d3e452d9b4f491530e2a0c03cb/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 2966519bd31da85fe6f42591426185fc21bd99d3e452d9b4f491530e2a0c03cb -s
1 S root       29453       1   29453  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/7b79ed5657dc2a474654ac9a84e8709ab95d24dbfae11cd9f638e0ccdfe3e0b7/userdata -c 7b79ed5657dc2a474654ac9a84e8709ab95d24dbfae11cd9f638e0ccdfe3e0b7 --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_metallb-operator-webhook-server-86c9c5f47f-wwgnb_a569ffff-2a47-4dfc-b245-9453c87786bf/webhook-server/0.log --log-level info -n k8s_webhook-server_metallb-operator-webhook-server-86c9c5f47f-wwgnb_metallb-system_a569ffff-2a47-4dfc-b245-9453c87786bf_0 -P /run/containers/storage/overlay-containers/7b79ed5657dc2a474654ac9a84e8709ab95d24dbfae11cd9f638e0ccdfe3e0b7/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/7b79ed5657dc2a474654ac9a84e8709ab95d24dbfae11cd9f638e0ccdfe3e0b7/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/7b79ed5657dc2a474654ac9a84e8709ab95d24dbfae11cd9f638e0ccdfe3e0b7/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 7b79ed5657dc2a474654ac9a84e8709ab95d24dbfae11cd9f638e0ccdfe3e0b7 -s
4 S 1000690+   29455   29451   29455  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:00 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   29466  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:03 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   29467  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:01 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   29468  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:00 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   29469  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:00 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   29470  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:00 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   29472  0   17  80   0 - 1668709 ep_pol 07:10 ?      00:00:01 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   29473  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:00 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   29474  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:00 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   29475  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:01 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   29482  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:00 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   29481  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:01 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   29489  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:00 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   30105  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:00 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   30107  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:01 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   30108  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:00 /manager --enable-leader-election --disable-cert-rotation=true
1 S 1000690+   29455   29451   30109  0   17  80   0 - 1668709 futex_ 07:10 ?      00:00:01 /manager --enable-leader-election --disable-cert-rotation=true
4 S 1000690+   29456   29453   29456  0   18  80   0 - 633023 futex_ 07:10 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   29476  0   18  80   0 - 633023 futex_ 07:10 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   29477  0   18  80   0 - 633023 futex_ 07:10 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   29478  0   18  80   0 - 633023 futex_ 07:10 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   29479  0   18  80   0 - 633023 futex_ 07:10 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   29480  0   18  80   0 - 633023 futex_ 07:10 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   29483  0   18  80   0 - 633023 ep_pol 07:10 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   29484  0   18  80   0 - 633023 futex_ 07:10 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   29485  0   18  80   0 - 633023 futex_ 07:10 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   29486  0   18  80   0 - 633023 futex_ 07:10 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   29487  0   18  80   0 - 633023 futex_ 07:10 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   29488  0   18  80   0 - 633023 futex_ 07:10 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   31392  0   18  80   0 - 633023 futex_ 07:11 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   32262  0   18  80   0 - 633023 futex_ 07:11 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   32263  0   18  80   0 - 633023 futex_ 07:11 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   32264  0   18  80   0 - 633023 futex_ 07:11 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   35564  0   18  80   0 - 633023 futex_ 07:13 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S 1000690+   29456   29453   82230  0   18  80   0 - 633023 futex_ 07:28 ?       00:00:00 /controller --disable-cert-rotation=true --port=7472 --log-level=info --webhook-mode=onlywebhook
1 S root       30136       1   30136  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/1dec21ccc521244df1d2b6ff8e807902ecd0712aa7be075e696922c383a66f3a/userdata -c 1dec21ccc521244df1d2b6ff8e807902ecd0712aa7be075e696922c383a66f3a --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_frr-k8s-h7lfx_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9/1dec21ccc521244df1d2b6ff8e807902ecd0712aa7be075e696922c383a66f3a.log --log-level info -n k8s_POD_frr-k8s-h7lfx_metallb-system_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9_0 -P /run/containers/storage/overlay-containers/1dec21ccc521244df1d2b6ff8e807902ecd0712aa7be075e696922c383a66f3a/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/1dec21ccc521244df1d2b6ff8e807902ecd0712aa7be075e696922c383a66f3a/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/1dec21ccc521244df1d2b6ff8e807902ecd0712aa7be075e696922c383a66f3a/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 1dec21ccc521244df1d2b6ff8e807902ecd0712aa7be075e696922c383a66f3a -s
4 S root       30139   30136   30139  0    1  80   0 -   624 do_sys 07:10 ?        00:00:00 /usr/bin/pod
1 S root       30198       1   30198  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/076ca752a874a45a3b96ccdf36258eba06e855e0a4c07236bff5ccfea72053e9/userdata -c 076ca752a874a45a3b96ccdf36258eba06e855e0a4c07236bff5ccfea72053e9 --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_controller-68d546b9d8-hvl4d_882eba24-ec61-49fd-8048-32ded4a05a45/controller/0.log --log-level info -n k8s_controller_controller-68d546b9d8-hvl4d_metallb-system_882eba24-ec61-49fd-8048-32ded4a05a45_0 -P /run/containers/storage/overlay-containers/076ca752a874a45a3b96ccdf36258eba06e855e0a4c07236bff5ccfea72053e9/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/076ca752a874a45a3b96ccdf36258eba06e855e0a4c07236bff5ccfea72053e9/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/076ca752a874a45a3b96ccdf36258eba06e855e0a4c07236bff5ccfea72053e9/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 076ca752a874a45a3b96ccdf36258eba06e855e0a4c07236bff5ccfea72053e9 -s
4 S 1000690+   30200   30198   30200  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   30256  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:01 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   30257  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   30258  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   30259  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   30260  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   30262  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
5 S 1000690+   30200   30198   30263  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   30264  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   30268  0   17  80   0 - 614654 ep_pol 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   30269  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   30270  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   30271  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   30272  0   17  80   0 - 614654 futex_ 07:10 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   40905  0   17  80   0 - 614654 futex_ 07:15 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   87237  0   17  80   0 - 614654 futex_ 07:33 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S 1000690+   30200   30198   87238  0   17  80   0 - 614654 futex_ 07:33 ?       00:00:00 /controller --port=29150 --log-level=debug --webhook-mode=disabled --tls-min-version=VersionTLS12
1 S root       30297       1   30297  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/55bd379c459eebda29999c419f30d06a09a7377c8162e032ad6f41b77dbb491c/userdata -c 55bd379c459eebda29999c419f30d06a09a7377c8162e032ad6f41b77dbb491c --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_controller-68d546b9d8-hvl4d_882eba24-ec61-49fd-8048-32ded4a05a45/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_controller-68d546b9d8-hvl4d_metallb-system_882eba24-ec61-49fd-8048-32ded4a05a45_0 -P /run/containers/storage/overlay-containers/55bd379c459eebda29999c419f30d06a09a7377c8162e032ad6f41b77dbb491c/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/55bd379c459eebda29999c419f30d06a09a7377c8162e032ad6f41b77dbb491c/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/55bd379c459eebda29999c419f30d06a09a7377c8162e032ad6f41b77dbb491c/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 55bd379c459eebda29999c419f30d06a09a7377c8162e032ad6f41b77dbb491c -s
4 S 1000690+   30299   30297   30299  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://127.0.0.1:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S 1000690+   30299   30297   30303  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://127.0.0.1:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S 1000690+   30299   30297   30304  0   11  80   0 - 503797 ep_pol 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://127.0.0.1:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S 1000690+   30299   30297   30305  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://127.0.0.1:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S 1000690+   30299   30297   30306  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://127.0.0.1:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S 1000690+   30299   30297   30307  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://127.0.0.1:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S 1000690+   30299   30297   30308  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://127.0.0.1:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S 1000690+   30299   30297   30309  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://127.0.0.1:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S 1000690+   30299   30297   30310  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://127.0.0.1:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S 1000690+   30299   30297   30311  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://127.0.0.1:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S 1000690+   30299   30297   30312  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://127.0.0.1:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S root       30320       1   30320  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/026c5e43e0d4741b0a295325a2d96e5bfa1d135d8f1d273b52ff840d2b06e26f/userdata -c 026c5e43e0d4741b0a295325a2d96e5bfa1d135d8f1d273b52ff840d2b06e26f --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_speaker-ln5cj_66aee647-ecf6-432b-95a9-6f2bcf7cf9cf/speaker/0.log --log-level info -n k8s_speaker_speaker-ln5cj_metallb-system_66aee647-ecf6-432b-95a9-6f2bcf7cf9cf_0 -P /run/containers/storage/overlay-containers/026c5e43e0d4741b0a295325a2d96e5bfa1d135d8f1d273b52ff840d2b06e26f/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/026c5e43e0d4741b0a295325a2d96e5bfa1d135d8f1d273b52ff840d2b06e26f/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/026c5e43e0d4741b0a295325a2d96e5bfa1d135d8f1d273b52ff840d2b06e26f/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 026c5e43e0d4741b0a295325a2d96e5bfa1d135d8f1d273b52ff840d2b06e26f -s
4 S root       30322   30320   30322  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:05 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30322   30320   30326  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:25 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30322   30320   30327  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:05 /speaker --port=29150 --log-level=debug --host=localhost
5 S root       30322   30320   30328  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:05 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30322   30320   30329  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:00 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30322   30320   30330  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:00 /speaker --port=29150 --log-level=debug --host=localhost
5 S root       30322   30320   30331  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:04 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30322   30320   30332  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:05 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30322   30320   30333  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:05 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30322   30320   30334  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:05 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30322   30320   30335  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:00 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30322   30320   30336  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:04 /speaker --port=29150 --log-level=debug --host=localhost
5 S root       30322   30320   30337  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:04 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30322   30320   30338  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:04 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30322   30320   30339  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:05 /speaker --port=29150 --log-level=debug --host=localhost
5 S root       30322   30320   30340  0   18  80   0 - 633968 futex_ 07:10 ?       00:00:05 /speaker --port=29150 --log-level=debug --host=localhost
5 S root       30322   30320   32580  0   18  80   0 - 633968 futex_ 07:12 ?       00:00:05 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30322   30320   77228  0   18  80   0 - 633968 ep_pol 07:23 ?       00:00:05 /speaker --port=29150 --log-level=debug --host=localhost
1 S root       30343       1   30343  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/d4f88029bc7d091531715505e60520eca061bb84100da3442a335fbc0d4e3956/userdata -c d4f88029bc7d091531715505e60520eca061bb84100da3442a335fbc0d4e3956 --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_speaker-ln5cj_66aee647-ecf6-432b-95a9-6f2bcf7cf9cf/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_speaker-ln5cj_metallb-system_66aee647-ecf6-432b-95a9-6f2bcf7cf9cf_0 -P /run/containers/storage/overlay-containers/d4f88029bc7d091531715505e60520eca061bb84100da3442a335fbc0d4e3956/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/d4f88029bc7d091531715505e60520eca061bb84100da3442a335fbc0d4e3956/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/d4f88029bc7d091531715505e60520eca061bb84100da3442a335fbc0d4e3956/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u d4f88029bc7d091531715505e60520eca061bb84100da3442a335fbc0d4e3956 -s
4 S nfsnobo+   30348   30343   30348  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://localhost:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30348   30343   30352  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://localhost:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30348   30343   30353  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://localhost:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30348   30343   30354  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://localhost:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30348   30343   30355  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://localhost:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30348   30343   30356  0   11  80   0 - 503797 ep_pol 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://localhost:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30348   30343   30357  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://localhost:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30348   30343   30358  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://localhost:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30348   30343   30359  0   11  80   0 - 503797 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://localhost:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30348   30343   33727  0   11  80   0 - 503797 futex_ 07:12 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://localhost:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30348   30343  121278  0   11  80   0 - 503797 futex_ 08:06 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9120 --upstream=http://localhost:29150/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S root       30439       1   30439  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/4f31e7a523f938f61ef0267c445095a611ed11b542b9f6a8af5a59343ae9b7b7/userdata -c 4f31e7a523f938f61ef0267c445095a611ed11b542b9f6a8af5a59343ae9b7b7 --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-zv2bz_d3603911-89b0-4e7e-bec3-f7496980b97a/frr-k8s-webhook-server/0.log --log-level info -n k8s_frr-k8s-webhook-server_frr-k8s-webhook-server-64bf5d555-zv2bz_metallb-system_d3603911-89b0-4e7e-bec3-f7496980b97a_0 -P /run/containers/storage/overlay-containers/4f31e7a523f938f61ef0267c445095a611ed11b542b9f6a8af5a59343ae9b7b7/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/4f31e7a523f938f61ef0267c445095a611ed11b542b9f6a8af5a59343ae9b7b7/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/4f31e7a523f938f61ef0267c445095a611ed11b542b9f6a8af5a59343ae9b7b7/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 4f31e7a523f938f61ef0267c445095a611ed11b542b9f6a8af5a59343ae9b7b7 -s
4 S 1000690+   30441   30439   30441  0   18  80   0 - 631254 futex_ 07:10 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   30448  0   18  80   0 - 631254 futex_ 07:10 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   30449  0   18  80   0 - 631254 futex_ 07:10 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   30450  0   18  80   0 - 631254 futex_ 07:10 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
5 S 1000690+   30441   30439   30451  0   18  80   0 - 631254 futex_ 07:10 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   30452  0   18  80   0 - 631254 futex_ 07:10 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   30453  0   18  80   0 - 631254 ep_pol 07:10 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   30456  0   18  80   0 - 631254 futex_ 07:10 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   30457  0   18  80   0 - 631254 futex_ 07:10 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   30458  0   18  80   0 - 631254 futex_ 07:10 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   30459  0   18  80   0 - 631254 futex_ 07:10 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   32299  0   18  80   0 - 631254 futex_ 07:11 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   32300  0   18  80   0 - 631254 futex_ 07:11 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   32301  0   18  80   0 - 631254 futex_ 07:11 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   46333  0   18  80   0 - 631254 futex_ 07:16 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   79710  0   18  80   0 - 631254 futex_ 07:26 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439   79711  0   18  80   0 - 631254 futex_ 07:26 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S 1000690+   30441   30439  137857  0   18  80   0 - 631254 futex_ 08:21 ?       00:00:00 /frr-k8s --log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=metallb-system --metrics-bind-address=:7572
1 S root       30546       1   30546  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/80ac98a25802020822242d7607a810f61afc66e00bb360412ee6a6141e864c82/userdata -c 80ac98a25802020822242d7607a810f61afc66e00bb360412ee6a6141e864c82 --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_frr-k8s-h7lfx_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9/controller/0.log --log-level info -n k8s_controller_frr-k8s-h7lfx_metallb-system_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9_0 -P /run/containers/storage/overlay-containers/80ac98a25802020822242d7607a810f61afc66e00bb360412ee6a6141e864c82/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/80ac98a25802020822242d7607a810f61afc66e00bb360412ee6a6141e864c82/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/80ac98a25802020822242d7607a810f61afc66e00bb360412ee6a6141e864c82/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 80ac98a25802020822242d7607a810f61afc66e00bb360412ee6a6141e864c82 -s
4 S root       30549   30546   30549  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   30553  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:01 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
5 S root       30549   30546   30554  0   17  80   0 - 612885 ep_pol 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   30555  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   30556  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   30557  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   30558  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   30559  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   30560  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   30561  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   30562  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   30563  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   30564  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   30565  0   17  80   0 - 612885 futex_ 07:10 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   33682  0   17  80   0 - 612885 futex_ 07:12 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   79898  0   17  80   0 - 612885 futex_ 07:26 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30549   30546   79899  0   17  80   0 - 612885 futex_ 07:26 ?       00:00:00 /frr-k8s --node-name=crc --namespace=metallb-system --metrics-bind-address=127.0.0.1:7572 --log-level=debug
1 S root       30567       1   30567  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e5cb4cfcff6660cb9d725f5c2763107afd826c8dc231388b47ca4f5c91a8d8de/userdata -c e5cb4cfcff6660cb9d725f5c2763107afd826c8dc231388b47ca4f5c91a8d8de --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_frr-k8s-h7lfx_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9/frr/0.log --log-level info -n k8s_frr_frr-k8s-h7lfx_metallb-system_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9_0 -P /run/containers/storage/overlay-containers/e5cb4cfcff6660cb9d725f5c2763107afd826c8dc231388b47ca4f5c91a8d8de/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e5cb4cfcff6660cb9d725f5c2763107afd826c8dc231388b47ca4f5c91a8d8de/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e5cb4cfcff6660cb9d725f5c2763107afd826c8dc231388b47ca4f5c91a8d8de/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e5cb4cfcff6660cb9d725f5c2763107afd826c8dc231388b47ca4f5c91a8d8de -s
4 S root       30570   30567   30570  0    1  80   0 -  1233 do_pol 07:10 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=tail /usr/bin/tail -f /etc/frr/frr.log
4 S root       30574   30570   30574  0    1  80   0 -   271 signal 07:10 ?        00:00:00 /sbin/tini -- /usr/lib/frr/docker-start
4 S root       30576   30574   30576  0    1  80   0 -  1113 do_wai 07:10 ?        00:00:00 /bin/bash /usr/lib/frr/docker-start
4 S root       30580   30576   30580  0    1  80   0 - 11292 do_pol 07:10 ?        00:00:01 /usr/lib/frr/watchfrr zebra bgpd staticd bfdd
1 S root       30595       1   30595  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/68dfc9c5045f8ba2a75017fedf391571c078d600407f6cb29a5f33bf775ad486/userdata -c 68dfc9c5045f8ba2a75017fedf391571c078d600407f6cb29a5f33bf775ad486 --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_frr-k8s-h7lfx_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9/reloader/0.log --log-level info -n k8s_reloader_frr-k8s-h7lfx_metallb-system_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9_0 -P /run/containers/storage/overlay-containers/68dfc9c5045f8ba2a75017fedf391571c078d600407f6cb29a5f33bf775ad486/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/68dfc9c5045f8ba2a75017fedf391571c078d600407f6cb29a5f33bf775ad486/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/68dfc9c5045f8ba2a75017fedf391571c078d600407f6cb29a5f33bf775ad486/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 68dfc9c5045f8ba2a75017fedf391571c078d600407f6cb29a5f33bf775ad486 -s
4 S root       30598   30595   30598  0    1  80   0 -  1047 do_wai 07:10 ?        00:00:00 /bin/bash /etc/frr_reloader/frr-reloader.sh
5 S etcd       30599   30574   30599  0    8  80   0 - 191635 do_pol 07:10 ?       00:00:02 /usr/libexec/frr/zebra -d -F traditional -A 127.0.0.1 -s 90000000
1 S etcd       30599   30574   30600  0    8  80   0 - 191635 futex_ 07:10 ?       00:00:00 /usr/libexec/frr/zebra -d -F traditional -A 127.0.0.1 -s 90000000
5 S etcd       30599   30574   30601  0    8  80   0 - 191635 do_pol 07:10 ?       00:00:00 /usr/libexec/frr/zebra -d -F traditional -A 127.0.0.1 -s 90000000
5 S etcd       30599   30574   30602  0    8  80   0 - 191635 do_pol 07:10 ?       00:00:00 /usr/libexec/frr/zebra -d -F traditional -A 127.0.0.1 -s 90000000
5 S etcd       30599   30574   30615  0    8  80   0 - 191635 do_pol 07:10 ?       00:00:00 /usr/libexec/frr/zebra -d -F traditional -A 127.0.0.1 -s 90000000
5 S etcd       30599   30574   30616  0    8  80   0 - 191635 do_pol 07:10 ?       00:00:00 /usr/libexec/frr/zebra -d -F traditional -A 127.0.0.1 -s 90000000
5 S etcd       30599   30574   30619  0    8  80   0 - 191635 do_pol 07:10 ?       00:00:00 /usr/libexec/frr/zebra -d -F traditional -A 127.0.0.1 -s 90000000
5 S etcd       30599   30574   30623  0    8  80   0 - 191635 do_pol 07:10 ?       00:00:00 /usr/libexec/frr/zebra -d -F traditional -A 127.0.0.1 -s 90000000
5 S etcd       30611   30574   30611  0    4  80   0 - 85912 do_pol 07:10 ?        00:00:01 /usr/libexec/frr/bgpd -d -F traditional -A 127.0.0.1 -p 0
1 S etcd       30611   30574   30612  0    4  80   0 - 85912 futex_ 07:10 ?        00:00:00 /usr/libexec/frr/bgpd -d -F traditional -A 127.0.0.1 -p 0
5 S etcd       30611   30574   30613  0    4  80   0 - 85912 do_pol 07:10 ?        00:00:00 /usr/libexec/frr/bgpd -d -F traditional -A 127.0.0.1 -p 0
1 S etcd       30611   30574   30614  0    4  80   0 - 85912 futex_ 07:10 ?        00:00:00 /usr/libexec/frr/bgpd -d -F traditional -A 127.0.0.1 -p 0
5 S etcd       30618   30574   30618  0    1  80   0 - 12130 do_pol 07:10 ?        00:00:00 /usr/libexec/frr/staticd -d -F traditional -A 127.0.0.1
5 S etcd       30622   30574   30622  0    1  80   0 - 12161 do_pol 07:10 ?        00:00:00 /usr/libexec/frr/bfdd -d -F traditional -A 127.0.0.1
1 S root       30625       1   30625  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/413a4cccd6f8db63dc4c0e03ef16e87f07b6ef9719f9c5ab6021bf5f282c5a02/userdata -c 413a4cccd6f8db63dc4c0e03ef16e87f07b6ef9719f9c5ab6021bf5f282c5a02 --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_frr-k8s-h7lfx_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9/frr-metrics/0.log --log-level info -n k8s_frr-metrics_frr-k8s-h7lfx_metallb-system_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9_0 -P /run/containers/storage/overlay-containers/413a4cccd6f8db63dc4c0e03ef16e87f07b6ef9719f9c5ab6021bf5f282c5a02/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/413a4cccd6f8db63dc4c0e03ef16e87f07b6ef9719f9c5ab6021bf5f282c5a02/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/413a4cccd6f8db63dc4c0e03ef16e87f07b6ef9719f9c5ab6021bf5f282c5a02/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 413a4cccd6f8db63dc4c0e03ef16e87f07b6ef9719f9c5ab6021bf5f282c5a02 -s
4 S root       30628   30625   30628  0   11  80   0 - 500626 futex_ 07:10 ?       00:00:00 /etc/frr_metrics/frr-metrics --metrics-port=7573 --metrics-bind-address=127.0.0.1
1 S root       30628   30625   30632  0   11  80   0 - 500626 futex_ 07:10 ?       00:00:01 /etc/frr_metrics/frr-metrics --metrics-port=7573 --metrics-bind-address=127.0.0.1
1 S root       30628   30625   30633  0   11  80   0 - 500626 futex_ 07:10 ?       00:00:00 /etc/frr_metrics/frr-metrics --metrics-port=7573 --metrics-bind-address=127.0.0.1
1 S root       30628   30625   30634  0   11  80   0 - 500626 futex_ 07:10 ?       00:00:00 /etc/frr_metrics/frr-metrics --metrics-port=7573 --metrics-bind-address=127.0.0.1
1 S root       30628   30625   30635  0   11  80   0 - 500626 ep_pol 07:10 ?       00:00:00 /etc/frr_metrics/frr-metrics --metrics-port=7573 --metrics-bind-address=127.0.0.1
1 S root       30628   30625   30636  0   11  80   0 - 500626 futex_ 07:10 ?       00:00:00 /etc/frr_metrics/frr-metrics --metrics-port=7573 --metrics-bind-address=127.0.0.1
1 S root       30628   30625   30637  0   11  80   0 - 500626 futex_ 07:10 ?       00:00:00 /etc/frr_metrics/frr-metrics --metrics-port=7573 --metrics-bind-address=127.0.0.1
1 S root       30628   30625   30638  0   11  80   0 - 500626 futex_ 07:10 ?       00:00:00 /etc/frr_metrics/frr-metrics --metrics-port=7573 --metrics-bind-address=127.0.0.1
1 S root       30628   30625   30639  0   11  80   0 - 500626 futex_ 07:10 ?       00:00:00 /etc/frr_metrics/frr-metrics --metrics-port=7573 --metrics-bind-address=127.0.0.1
1 S root       30628   30625   30640  0   11  80   0 - 500626 futex_ 07:10 ?       00:00:00 /etc/frr_metrics/frr-metrics --metrics-port=7573 --metrics-bind-address=127.0.0.1
1 S root       30628   30625   30641  0   11  80   0 - 500626 futex_ 07:10 ?       00:00:00 /etc/frr_metrics/frr-metrics --metrics-port=7573 --metrics-bind-address=127.0.0.1
1 S root       30643       1   30643  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/668a8e48c413bfbbf79c3ec73f17633e80318e6cbf4abc7bbc919daca303e111/userdata -c 668a8e48c413bfbbf79c3ec73f17633e80318e6cbf4abc7bbc919daca303e111 --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_frr-k8s-h7lfx_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_frr-k8s-h7lfx_metallb-system_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9_0 -P /run/containers/storage/overlay-containers/668a8e48c413bfbbf79c3ec73f17633e80318e6cbf4abc7bbc919daca303e111/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/668a8e48c413bfbbf79c3ec73f17633e80318e6cbf4abc7bbc919daca303e111/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/668a8e48c413bfbbf79c3ec73f17633e80318e6cbf4abc7bbc919daca303e111/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 668a8e48c413bfbbf79c3ec73f17633e80318e6cbf4abc7bbc919daca303e111 -s
4 S nfsnobo+   30646   30643   30646  0   11  80   0 - 503925 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9140 --upstream=http://127.0.0.1:7572/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30646   30643   30650  0   11  80   0 - 503925 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9140 --upstream=http://127.0.0.1:7572/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30646   30643   30651  0   11  80   0 - 503925 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9140 --upstream=http://127.0.0.1:7572/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30646   30643   30652  0   11  80   0 - 503925 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9140 --upstream=http://127.0.0.1:7572/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30646   30643   30653  0   11  80   0 - 503925 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9140 --upstream=http://127.0.0.1:7572/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30646   30643   30654  0   11  80   0 - 503925 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9140 --upstream=http://127.0.0.1:7572/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30646   30643   30655  0   11  80   0 - 503925 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9140 --upstream=http://127.0.0.1:7572/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30646   30643   30656  0   11  80   0 - 503925 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9140 --upstream=http://127.0.0.1:7572/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30646   30643   30657  0   11  80   0 - 503925 ep_pol 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9140 --upstream=http://127.0.0.1:7572/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30646   30643   30658  0   11  80   0 - 503925 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9140 --upstream=http://127.0.0.1:7572/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30646   30643   30659  0   11  80   0 - 503925 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9140 --upstream=http://127.0.0.1:7572/ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S root       30661       1   30661  0    1  80   0 -  2078 do_pol 07:10 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e9260212ded6a0efe24d446f8b95e94eff9e01136aee42c9f68aebc165b74f2c/userdata -c e9260212ded6a0efe24d446f8b95e94eff9e01136aee42c9f68aebc165b74f2c --exit-dir /var/run/crio/exits -l /var/log/pods/metallb-system_frr-k8s-h7lfx_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9/kube-rbac-proxy-frr/0.log --log-level info -n k8s_kube-rbac-proxy-frr_frr-k8s-h7lfx_metallb-system_e16cedf7-6d0e-4161-9ffb-1ba8ad4eaee9_0 -P /run/containers/storage/overlay-containers/e9260212ded6a0efe24d446f8b95e94eff9e01136aee42c9f68aebc165b74f2c/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e9260212ded6a0efe24d446f8b95e94eff9e01136aee42c9f68aebc165b74f2c/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e9260212ded6a0efe24d446f8b95e94eff9e01136aee42c9f68aebc165b74f2c/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e9260212ded6a0efe24d446f8b95e94eff9e01136aee42c9f68aebc165b74f2c -s
4 S nfsnobo+   30664   30661   30664  0   12  80   0 - 522358 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9141 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:7573/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30664   30661   30668  0   12  80   0 - 522358 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9141 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:7573/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30664   30661   30669  0   12  80   0 - 522358 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9141 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:7573/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30664   30661   30670  0   12  80   0 - 522358 ep_pol 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9141 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:7573/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30664   30661   30671  0   12  80   0 - 522358 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9141 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:7573/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30664   30661   30672  0   12  80   0 - 522358 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9141 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:7573/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30664   30661   30673  0   12  80   0 - 522358 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9141 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:7573/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30664   30661   30674  0   12  80   0 - 522358 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9141 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:7573/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30664   30661   30675  0   12  80   0 - 522358 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9141 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:7573/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30664   30661   30676  0   12  80   0 - 522358 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9141 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:7573/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30664   30661   30677  0   12  80   0 - 522358 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9141 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:7573/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
1 S nfsnobo+   30664   30661   30678  0   12  80   0 - 522358 futex_ 07:10 ?       00:00:00 /usr/bin/kube-rbac-proxy --logtostderr --secure-listen-address=:9141 --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 --upstream=http://127.0.0.1:7573/ --tls-private-key-file=/etc/metrics/tls.key --tls-cert-file=/etc/metrics/tls.crt
0 S root       30798   30598   30798  0    1  80   0 -  1233 do_sys 07:10 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep infinity
1 S root       31305       1   31305  0    1  80   0 -  2078 do_pol 07:11 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/82ea83ae92387aa4c09a64f0a1ad9f090f0567a36ca1eacce570c23c963dd65d/userdata -c 82ea83ae92387aa4c09a64f0a1ad9f090f0567a36ca1eacce570c23c963dd65d --exit-dir /var/run/crio/exits -l /var/log/pods/cert-manager-operator_cert-manager-operator-controller-manager-57cd46d6d-bdwck_94e9adea-df96-46c6-86ac-5422f880b631/cert-manager-operator/0.log --log-level info -n k8s_cert-manager-operator_cert-manager-operator-controller-manager-57cd46d6d-bdwck_cert-manager-operator_94e9adea-df96-46c6-86ac-5422f880b631_0 -P /run/containers/storage/overlay-containers/82ea83ae92387aa4c09a64f0a1ad9f090f0567a36ca1eacce570c23c963dd65d/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/82ea83ae92387aa4c09a64f0a1ad9f090f0567a36ca1eacce570c23c963dd65d/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/82ea83ae92387aa4c09a64f0a1ad9f090f0567a36ca1eacce570c23c963dd65d/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 82ea83ae92387aa4c09a64f0a1ad9f090f0567a36ca1eacce570c23c963dd65d -s
4 S 1000700+   31307   31305   31307  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:00 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S 1000700+   31307   31305   31312  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:08 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
5 S 1000700+   31307   31305   31313  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:03 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S 1000700+   31307   31305   31314  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:01 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S 1000700+   31307   31305   31315  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:00 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S 1000700+   31307   31305   31316  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:00 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S 1000700+   31307   31305   31317  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:02 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S 1000700+   31307   31305   31318  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:02 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S 1000700+   31307   31305   31319  0   17  80   0 - 624130 ep_pol 07:11 ?       00:00:02 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
5 S 1000700+   31307   31305   31320  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:01 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
5 S 1000700+   31307   31305   31321  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:03 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S 1000700+   31307   31305   31322  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:03 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S 1000700+   31307   31305   31323  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:02 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
5 S 1000700+   31307   31305   31324  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:02 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S 1000700+   31307   31305   31329  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:01 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S 1000700+   31307   31305   31330  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:02 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S 1000700+   31307   31305   31615  0   17  80   0 - 624130 futex_ 07:11 ?       00:00:03 /usr/bin/cert-manager-operator start --v=2 --trusted-ca-configmap= --cloud-credentials-secret= --unsupported-addon-features=
1 S root       31519       1   31519  0    1  80   0 -  2078 do_pol 07:11 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/fe4aae2fbc1586d0b441481e548985cca0e2a42905c853670c90475219cf658d/userdata -c fe4aae2fbc1586d0b441481e548985cca0e2a42905c853670c90475219cf658d --exit-dir /var/run/crio/exits -l /var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-4w69l_96ab5737-7fd2-43bf-8f7c-b5b10990c2b5/cert-manager-cainjector/0.log --log-level info -n k8s_cert-manager-cainjector_cert-manager-cainjector-7d9f95dbf-4w69l_cert-manager_96ab5737-7fd2-43bf-8f7c-b5b10990c2b5_0 -P /run/containers/storage/overlay-containers/fe4aae2fbc1586d0b441481e548985cca0e2a42905c853670c90475219cf658d/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/fe4aae2fbc1586d0b441481e548985cca0e2a42905c853670c90475219cf658d/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/fe4aae2fbc1586d0b441481e548985cca0e2a42905c853670c90475219cf658d/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u fe4aae2fbc1586d0b441481e548985cca0e2a42905c853670c90475219cf658d -s
1 S root       31521       1   31521  0    1  80   0 -  2078 do_pol 07:11 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/0f87d96e852aa26226691f770a38f84f88eeaef975a659a7ac73ae35807d6170/userdata -c 0f87d96e852aa26226691f770a38f84f88eeaef975a659a7ac73ae35807d6170 --exit-dir /var/run/crio/exits -l /var/log/pods/cert-manager_cert-manager-webhook-d969966f-4lszr_eb52d59d-d019-4770-94cf-212c63c37fc6/cert-manager-webhook/0.log --log-level info -n k8s_cert-manager-webhook_cert-manager-webhook-d969966f-4lszr_cert-manager_eb52d59d-d019-4770-94cf-212c63c37fc6_0 -P /run/containers/storage/overlay-containers/0f87d96e852aa26226691f770a38f84f88eeaef975a659a7ac73ae35807d6170/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/0f87d96e852aa26226691f770a38f84f88eeaef975a659a7ac73ae35807d6170/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/0f87d96e852aa26226691f770a38f84f88eeaef975a659a7ac73ae35807d6170/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 0f87d96e852aa26226691f770a38f84f88eeaef975a659a7ac73ae35807d6170 -s
4 S 1000710+   31523   31519   31523  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:00 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31531  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:02 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31532  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:00 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
5 S 1000710+   31523   31519   31533  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:00 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31534  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:01 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31535  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:00 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31541  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:00 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31542  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:00 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31543  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:00 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31547  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:00 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31549  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:01 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31550  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:01 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31551  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:00 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31552  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:01 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31553  0   18  80   0 - 651653 ep_pol 07:11 ?       00:00:01 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31554  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:01 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31555  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:01 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
1 S 1000710+   31523   31519   31556  0   18  80   0 - 651653 futex_ 07:11 ?       00:00:00 /app/cmd/cainjector/cainjector --leader-election-namespace=kube-system --v=2
4 S 1000710+   31524   31521   31524  0   17  80   0 - 618485 futex_ 07:11 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   31536  0   17  80   0 - 618485 futex_ 07:11 ?       00:00:01 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   31537  0   17  80   0 - 618485 futex_ 07:11 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
5 S 1000710+   31524   31521   31538  0   17  80   0 - 618485 futex_ 07:11 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   31539  0   17  80   0 - 618485 futex_ 07:11 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   31540  0   17  80   0 - 618485 futex_ 07:11 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   31544  0   17  80   0 - 618485 futex_ 07:11 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   31545  0   17  80   0 - 618485 futex_ 07:11 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   31546  0   17  80   0 - 618485 futex_ 07:11 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   31548  0   17  80   0 - 618485 futex_ 07:11 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   34061  0   17  80   0 - 618485 futex_ 07:13 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   34527  0   17  80   0 - 618485 futex_ 07:13 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   37220  0   17  80   0 - 618485 futex_ 07:14 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   37221  0   17  80   0 - 618485 futex_ 07:14 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   37222  0   17  80   0 - 618485 futex_ 07:14 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   37954  0   17  80   0 - 618485 ep_pol 07:14 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S 1000710+   31524   31521   37955  0   17  80   0 - 618485 futex_ 07:14 ?       00:00:00 /app/cmd/webhook/webhook --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc --secure-port=10250 --v=2
1 S root       31927       1   31927  0    1  80   0 -  2078 do_pol 07:11 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/4be0a392d192d8d963a6eda9d6135222d798f2f2681f78a3e6c7a6b46a53861b/userdata -c 4be0a392d192d8d963a6eda9d6135222d798f2f2681f78a3e6c7a6b46a53861b --exit-dir /var/run/crio/exits -l /var/log/pods/cert-manager_cert-manager-7d4cc89fcb-l99g8_44861968-3c3b-4288-b1d2-0da6e2d74cc4/cert-manager-controller/0.log --log-level info -n k8s_cert-manager-controller_cert-manager-7d4cc89fcb-l99g8_cert-manager_44861968-3c3b-4288-b1d2-0da6e2d74cc4_0 -P /run/containers/storage/overlay-containers/4be0a392d192d8d963a6eda9d6135222d798f2f2681f78a3e6c7a6b46a53861b/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/4be0a392d192d8d963a6eda9d6135222d798f2f2681f78a3e6c7a6b46a53861b/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/4be0a392d192d8d963a6eda9d6135222d798f2f2681f78a3e6c7a6b46a53861b/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 4be0a392d192d8d963a6eda9d6135222d798f2f2681f78a3e6c7a6b46a53861b -s
4 S 1000710+   31929   31927   31929  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:00 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31933  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:02 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31934  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:02 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31935  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:01 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31936  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:02 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31937  0   17  80   0 - 621276 ep_pol 07:11 ?       00:00:04 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31938  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:00 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31939  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:00 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31940  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:00 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31941  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:03 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31942  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:06 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31943  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:03 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
5 S 1000710+   31929   31927   31944  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:03 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31945  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:04 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   31946  0   17  80   0 - 621276 futex_ 07:11 ?       00:00:02 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   32603  0   17  80   0 - 621276 futex_ 07:12 ?       00:00:03 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S 1000710+   31929   31927   32604  0   17  80   0 - 621276 futex_ 07:12 ?       00:00:04 /app/cmd/controller/controller --acme-http01-solver-image=registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --max-concurrent-challenges=60 --v=2
1 S root       32351       1   32351  0    1  80   0 -  2078 do_pol 07:11 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/62a8ce962fda6e5f1d052340807b567fa977ab275bac4aa37310a2c4e601afd2/userdata -c 62a8ce962fda6e5f1d052340807b567fa977ab275bac4aa37310a2c4e601afd2 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_openstack-operator-index-kkbw6_bd51530d-735f-4124-bb3e-4c3967f64a31/registry-server/0.log --log-level info -n k8s_registry-server_openstack-operator-index-kkbw6_openstack-operators_bd51530d-735f-4124-bb3e-4c3967f64a31_0 -P /run/containers/storage/overlay-containers/62a8ce962fda6e5f1d052340807b567fa977ab275bac4aa37310a2c4e601afd2/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/62a8ce962fda6e5f1d052340807b567fa977ab275bac4aa37310a2c4e601afd2/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/62a8ce962fda6e5f1d052340807b567fa977ab275bac4aa37310a2c4e601afd2/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 62a8ce962fda6e5f1d052340807b567fa977ab275bac4aa37310a2c4e601afd2 -s
4 S 1001       32354   32351   32354  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S 1001       32354   32351   32366  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S 1001       32354   32351   32367  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S 1001       32354   32351   32368  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S 1001       32354   32351   32369  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S 1001       32354   32351   32370  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
5 S 1001       32354   32351   32371  0   16  80   0 - 1386404 ep_pol 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S 1001       32354   32351   32372  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S 1001       32354   32351   32373  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S 1001       32354   32351   32374  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
5 S 1001       32354   32351   32375  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S 1001       32354   32351   32376  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
5 S 1001       32354   32351   32381  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S 1001       32354   32351   32382  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S 1001       32354   32351   32387  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S 1001       32354   32351   32388  0   16  80   0 - 1386404 futex_ 07:11 ?      00:00:00 /bin/opm serve /configs --cache-dir=/tmp/cache
1 S root       33166       1   33166  0    1  80   0 -  2078 do_pol 07:12 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e8bc4e947e4bb80e527534926128452251a68ba20f441e38e9507741e89a58e1/userdata -c e8bc4e947e4bb80e527534926128452251a68ba20f441e38e9507741e89a58e1 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_openstack-operator-controller-operator-86f8d7b75f-wl9x5_6526f954-e830-42ea-b0a1-6121ff0e32c1/operator/0.log --log-level info -n k8s_operator_openstack-operator-controller-operator-86f8d7b75f-wl9x5_openstack-operators_6526f954-e830-42ea-b0a1-6121ff0e32c1_0 -P /run/containers/storage/overlay-containers/e8bc4e947e4bb80e527534926128452251a68ba20f441e38e9507741e89a58e1/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e8bc4e947e4bb80e527534926128452251a68ba20f441e38e9507741e89a58e1/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e8bc4e947e4bb80e527534926128452251a68ba20f441e38e9507741e89a58e1/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e8bc4e947e4bb80e527534926128452251a68ba20f441e38e9507741e89a58e1 -s
4 S 1000660+   33168   33166   33168  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:00 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   33168   33166   33172  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:02 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   33168   33166   33173  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:00 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   33168   33166   33174  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:00 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   33168   33166   33175  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:00 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   33168   33166   33176  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:00 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   33168   33166   33177  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:01 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   33168   33166   33178  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:00 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   33168   33166   33179  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:00 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   33168   33166   33180  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:01 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   33168   33166   33181  0   18  80   0 - 633679 ep_pol 07:12 ?       00:00:01 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   33168   33166   33182  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:00 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   33168   33166   33183  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:01 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   33168   33166   33823  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:00 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   33168   33166   33824  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:01 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   33168   33166   33825  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:00 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   33168   33166   33826  0   18  80   0 - 633679 futex_ 07:12 ?       00:00:00 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   33168   33166   35461  0   18  80   0 - 633679 futex_ 07:13 ?       00:00:00 /operator --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S root       33454       1   33454  0    1  80   0 -  2078 do_pol 07:12 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/99df977339f261db7546f87cbe948ec30af2ede882ddba490ab8e01b82ba46ca/userdata -c 99df977339f261db7546f87cbe948ec30af2ede882ddba490ab8e01b82ba46ca --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_openstack-operator-controller-operator-86f8d7b75f-wl9x5_6526f954-e830-42ea-b0a1-6121ff0e32c1/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_openstack-operator-controller-operator-86f8d7b75f-wl9x5_openstack-operators_6526f954-e830-42ea-b0a1-6121ff0e32c1_0 -P /run/containers/storage/overlay-containers/99df977339f261db7546f87cbe948ec30af2ede882ddba490ab8e01b82ba46ca/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/99df977339f261db7546f87cbe948ec30af2ede882ddba490ab8e01b82ba46ca/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/99df977339f261db7546f87cbe948ec30af2ede882ddba490ab8e01b82ba46ca/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 99df977339f261db7546f87cbe948ec30af2ede882ddba490ab8e01b82ba46ca -s
4 S 1000660+   33456   33454   33456  0   11  80   0 - 318158 futex_ 07:12 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   33456   33454   33460  0   11  80   0 - 318158 futex_ 07:12 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   33456   33454   33461  0   11  80   0 - 318158 futex_ 07:12 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   33456   33454   33462  0   11  80   0 - 318158 futex_ 07:12 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   33456   33454   33463  0   11  80   0 - 318158 futex_ 07:12 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   33456   33454   33464  0   11  80   0 - 318158 futex_ 07:12 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   33456   33454   33465  0   11  80   0 - 318158 ep_pol 07:12 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   33456   33454   33466  0   11  80   0 - 318158 futex_ 07:12 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   33456   33454   33467  0   11  80   0 - 318158 futex_ 07:12 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   33456   33454   33468  0   11  80   0 - 318158 futex_ 07:12 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   33456   33454   33469  0   11  80   0 - 318158 futex_ 07:12 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S root       34864       1   34864  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/6ac37704a8ed6d2ba72eb850dbbf1094b4752efc9b546dbf47a4afea82e44b2f/userdata -c 6ac37704a8ed6d2ba72eb850dbbf1094b4752efc9b546dbf47a4afea82e44b2f --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_placement-operator-controller-manager-ccbfcb8c-tnrxp_b842fd5e-6125-40cf-b729-947e54309c87/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_placement-operator-controller-manager-ccbfcb8c-tnrxp_openstack-operators_b842fd5e-6125-40cf-b729-947e54309c87_0 -P /run/containers/storage/overlay-containers/6ac37704a8ed6d2ba72eb850dbbf1094b4752efc9b546dbf47a4afea82e44b2f/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/6ac37704a8ed6d2ba72eb850dbbf1094b4752efc9b546dbf47a4afea82e44b2f/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/6ac37704a8ed6d2ba72eb850dbbf1094b4752efc9b546dbf47a4afea82e44b2f/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 6ac37704a8ed6d2ba72eb850dbbf1094b4752efc9b546dbf47a4afea82e44b2f -s
1 S root       34869       1   34869  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/506aa87f5b8cbaa4a0296b93b5fc9cec4a61e80ac2287a8dd62eca229d10555b/userdata -c 506aa87f5b8cbaa4a0296b93b5fc9cec4a61e80ac2287a8dd62eca229d10555b --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-7dv7x_1a5d562f-6fe0-4bc9-a39a-7f7d8e9c1b63/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_swift-operator-controller-manager-76d5577b-7dv7x_openstack-operators_1a5d562f-6fe0-4bc9-a39a-7f7d8e9c1b63_0 -P /run/containers/storage/overlay-containers/506aa87f5b8cbaa4a0296b93b5fc9cec4a61e80ac2287a8dd62eca229d10555b/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/506aa87f5b8cbaa4a0296b93b5fc9cec4a61e80ac2287a8dd62eca229d10555b/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/506aa87f5b8cbaa4a0296b93b5fc9cec4a61e80ac2287a8dd62eca229d10555b/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 506aa87f5b8cbaa4a0296b93b5fc9cec4a61e80ac2287a8dd62eca229d10555b -s
1 S root       34871       1   34871  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/bffe11888c186d0d380a522d009511a1a5d071d6dfef9c48bab346913a16dde9/userdata -c bffe11888c186d0d380a522d009511a1a5d071d6dfef9c48bab346913a16dde9 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_telemetry-operator-controller-manager-5ffb97cddf-vf8h7_368ec6b5-ba12-467e-ab4c-d46a83c31483/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_telemetry-operator-controller-manager-5ffb97cddf-vf8h7_openstack-operators_368ec6b5-ba12-467e-ab4c-d46a83c31483_0 -P /run/containers/storage/overlay-containers/bffe11888c186d0d380a522d009511a1a5d071d6dfef9c48bab346913a16dde9/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/bffe11888c186d0d380a522d009511a1a5d071d6dfef9c48bab346913a16dde9/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/bffe11888c186d0d380a522d009511a1a5d071d6dfef9c48bab346913a16dde9/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u bffe11888c186d0d380a522d009511a1a5d071d6dfef9c48bab346913a16dde9 -s
4 S 65532      34873   34864   34873  0   11  80   0 - 318290 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34873   34864   34940  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34873   34864   34941  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34873   34864   34942  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34873   34864   34943  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34873   34864   34953  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34873   34864   34954  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34873   34864   34955  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34873   34864   34964  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34873   34864   34965  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34873   34864   35067  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 1000660+   34874   34869   34874  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   34874   34869   34936  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   34874   34869   34937  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   34874   34869   34938  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   34874   34869   34939  0   11  80   0 - 318290 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   34874   34869   34959  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   34874   34869   34960  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   34874   34869   34961  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   34874   34869   34962  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   34874   34869   34963  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   34874   34869   35043  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 65532      34875   34871   34875  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   34944  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   34945  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   34946  0   14  80   0 - 318354 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   34947  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   34948  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   34949  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   34950  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   34951  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   34952  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   34956  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   34957  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   34958  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34875   34871   35110  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S root       34914       1   34914  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/64fb4173d712273112a49271cc39bbdf2cde02030405bcf33ae1082be5384c8c/userdata -c 64fb4173d712273112a49271cc39bbdf2cde02030405bcf33ae1082be5384c8c --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_ovn-operator-controller-manager-855d7949fc-gnvp7_2514440b-b998-4078-834e-642c3bcae80f/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_ovn-operator-controller-manager-855d7949fc-gnvp7_openstack-operators_2514440b-b998-4078-834e-642c3bcae80f_0 -P /run/containers/storage/overlay-containers/64fb4173d712273112a49271cc39bbdf2cde02030405bcf33ae1082be5384c8c/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/64fb4173d712273112a49271cc39bbdf2cde02030405bcf33ae1082be5384c8c/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/64fb4173d712273112a49271cc39bbdf2cde02030405bcf33ae1082be5384c8c/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 64fb4173d712273112a49271cc39bbdf2cde02030405bcf33ae1082be5384c8c -s
1 S root       34915       1   34915  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/13282c4e33b3a2764cfca7099879eccd4971c2cfeb8c55ded1521e589820490d/userdata -c 13282c4e33b3a2764cfca7099879eccd4971c2cfeb8c55ded1521e589820490d --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_watcher-operator-controller-manager-5595cf6c95-4hcbj_9ab8a530-fefb-477b-85f7-6f716f684292/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_watcher-operator-controller-manager-5595cf6c95-4hcbj_openstack-operators_9ab8a530-fefb-477b-85f7-6f716f684292_0 -P /run/containers/storage/overlay-containers/13282c4e33b3a2764cfca7099879eccd4971c2cfeb8c55ded1521e589820490d/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/13282c4e33b3a2764cfca7099879eccd4971c2cfeb8c55ded1521e589820490d/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/13282c4e33b3a2764cfca7099879eccd4971c2cfeb8c55ded1521e589820490d/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 13282c4e33b3a2764cfca7099879eccd4971c2cfeb8c55ded1521e589820490d -s
1 S root       34918       1   34918  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:01 /usr/bin/conmon -b /run/containers/storage/overlay-containers/2c476da5e325977ec9a66d5a02534364ad61ddd819b334add6a5f166f13ce5ad/userdata -c 2c476da5e325977ec9a66d5a02534364ad61ddd819b334add6a5f166f13ce5ad --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_openstack-operator-controller-manager-54df7874c5-l442f_9ac05823-a077-4e2a-8bf8-b5e9a454bf03/manager/0.log --log-level info -n k8s_manager_openstack-operator-controller-manager-54df7874c5-l442f_openstack-operators_9ac05823-a077-4e2a-8bf8-b5e9a454bf03_0 -P /run/containers/storage/overlay-containers/2c476da5e325977ec9a66d5a02534364ad61ddd819b334add6a5f166f13ce5ad/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/2c476da5e325977ec9a66d5a02534364ad61ddd819b334add6a5f166f13ce5ad/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/2c476da5e325977ec9a66d5a02534364ad61ddd819b334add6a5f166f13ce5ad/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 2c476da5e325977ec9a66d5a02534364ad61ddd819b334add6a5f166f13ce5ad -s
4 S 65532      34920   34915   34920  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34920   34915   34974  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34920   34915   34975  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34920   34915   34976  0   13  80   0 - 318226 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34920   34915   34978  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34920   34915   34983  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34920   34915   35001  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34920   34915   35002  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34920   34915   35003  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34920   34915   35020  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34920   34915   35021  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34920   34915   35022  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34920   34915   35109  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 65532      34921   34914   34921  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34921   34914   34984  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34921   34914   34985  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34921   34914   34986  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34921   34914   34997  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34921   34914   34998  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34921   34914   34999  0   11  80   0 - 318290 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34921   34914   35017  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34921   34914   35018  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34921   34914   35019  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      34921   34914   35086  0   11  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 65532      34922   34918   34922  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   34991  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:08 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   34992  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:09 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   34993  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:09 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   34994  0   18  80   0 - 653933 ep_pol 07:13 ?       00:00:07 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   34995  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   35000  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:06 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   35004  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:06 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   35006  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:08 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      34922   34918   35011  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:07 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   35012  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   35039  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:07 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   35040  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:05 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   35041  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:07 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   35042  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:06 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   35053  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:07 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   35054  0   18  80   0 - 653933 futex_ 07:13 ?       00:00:07 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      34922   34918   42148  0   18  80   0 - 653933 futex_ 07:15 ?       00:00:08 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S root       35045       1   35045  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/c773209bbb1cfc2cb9046f41fdb2832c2dd1a79322e6471dc25e52074ab71044/userdata -c c773209bbb1cfc2cb9046f41fdb2832c2dd1a79322e6471dc25e52074ab71044 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_openstack-operator-controller-manager-54df7874c5-l442f_9ac05823-a077-4e2a-8bf8-b5e9a454bf03/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_openstack-operator-controller-manager-54df7874c5-l442f_openstack-operators_9ac05823-a077-4e2a-8bf8-b5e9a454bf03_0 -P /run/containers/storage/overlay-containers/c773209bbb1cfc2cb9046f41fdb2832c2dd1a79322e6471dc25e52074ab71044/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/c773209bbb1cfc2cb9046f41fdb2832c2dd1a79322e6471dc25e52074ab71044/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/c773209bbb1cfc2cb9046f41fdb2832c2dd1a79322e6471dc25e52074ab71044/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u c773209bbb1cfc2cb9046f41fdb2832c2dd1a79322e6471dc25e52074ab71044 -s
4 S 65532      35047   35045   35047  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35057  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35058  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35059  0   14  80   0 - 318354 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35060  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35061  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35062  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35063  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35064  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35065  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35066  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35126  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35127  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      35047   35045   35132  0   14  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S root       36002       1   36002  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/04b49ed15cceb5d4bd66f48d3cf11265a8b40d43bca8eaf01ec50cf9d8e11a54/userdata -c 04b49ed15cceb5d4bd66f48d3cf11265a8b40d43bca8eaf01ec50cf9d8e11a54 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_horizon-operator-controller-manager-586b66cf4f-qxv5t_84c853db-41b6-4013-b6a6-b9f6c3fa74e3/manager/0.log --log-level info -n k8s_manager_horizon-operator-controller-manager-586b66cf4f-qxv5t_openstack-operators_84c853db-41b6-4013-b6a6-b9f6c3fa74e3_0 -P /run/containers/storage/overlay-containers/04b49ed15cceb5d4bd66f48d3cf11265a8b40d43bca8eaf01ec50cf9d8e11a54/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/04b49ed15cceb5d4bd66f48d3cf11265a8b40d43bca8eaf01ec50cf9d8e11a54/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/04b49ed15cceb5d4bd66f48d3cf11265a8b40d43bca8eaf01ec50cf9d8e11a54/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 04b49ed15cceb5d4bd66f48d3cf11265a8b40d43bca8eaf01ec50cf9d8e11a54 -s
1 S root       36003       1   36003  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/5011e80d4c66d6c347d1e6c7b9e5cb552f10f58e08a485bd93bc8a4603370091/userdata -c 5011e80d4c66d6c347d1e6c7b9e5cb552f10f58e08a485bd93bc8a4603370091 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7fb4f565cd5xb48_aaf10555-293e-4c2b-baea-ffcdd4eeb046/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_openstack-baremetal-operator-controller-manager-7fb4f565cd5xb48_openstack-operators_aaf10555-293e-4c2b-baea-ffcdd4eeb046_0 -P /run/containers/storage/overlay-containers/5011e80d4c66d6c347d1e6c7b9e5cb552f10f58e08a485bd93bc8a4603370091/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/5011e80d4c66d6c347d1e6c7b9e5cb552f10f58e08a485bd93bc8a4603370091/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/5011e80d4c66d6c347d1e6c7b9e5cb552f10f58e08a485bd93bc8a4603370091/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 5011e80d4c66d6c347d1e6c7b9e5cb552f10f58e08a485bd93bc8a4603370091 -s
1 S root       36006       1   36006  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/53668c90968112f0b2ee732fd83415149fe4ba8bb415f5c442daffaf03ba555e/userdata -c 53668c90968112f0b2ee732fd83415149fe4ba8bb415f5c442daffaf03ba555e --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_test-operator-controller-manager-6bb6dcddc-hmtgk_d6c2b423-4398-416f-bb17-70a8eb814964/manager/0.log --log-level info -n k8s_manager_test-operator-controller-manager-6bb6dcddc-hmtgk_openstack-operators_d6c2b423-4398-416f-bb17-70a8eb814964_0 -P /run/containers/storage/overlay-containers/53668c90968112f0b2ee732fd83415149fe4ba8bb415f5c442daffaf03ba555e/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/53668c90968112f0b2ee732fd83415149fe4ba8bb415f5c442daffaf03ba555e/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/53668c90968112f0b2ee732fd83415149fe4ba8bb415f5c442daffaf03ba555e/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 53668c90968112f0b2ee732fd83415149fe4ba8bb415f5c442daffaf03ba555e -s
4 S 1000660+   36010   36003   36010  0   12  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   36010   36003   36063  0   12  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   36010   36003   36064  0   12  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   36010   36003   36065  0   12  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   36010   36003   36067  0   12  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   36010   36003   36073  0   12  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   36010   36003   36074  0   12  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   36010   36003   36086  0   12  80   0 - 318290 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   36010   36003   36088  0   12  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   36010   36003   36089  0   12  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   36010   36003   36091  0   12  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 1000660+   36010   36003   36358  0   12  80   0 - 318290 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 65532      36011   36006   36011  0   16  80   0 - 596647 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   36079  0   16  80   0 - 596647 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   36080  0   16  80   0 - 596647 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   36081  0   16  80   0 - 596647 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   36082  0   16  80   0 - 596647 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   36083  0   16  80   0 - 596647 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   36090  0   16  80   0 - 596647 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   36092  0   16  80   0 - 596647 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   36094  0   16  80   0 - 596647 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   36107  0   16  80   0 - 596647 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   36108  0   16  80   0 - 596647 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   36154  0   16  80   0 - 596647 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   37386  0   16  80   0 - 596647 futex_ 07:14 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   37387  0   16  80   0 - 596647 futex_ 07:14 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   37388  0   16  80   0 - 596647 futex_ 07:14 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36011   36006   37389  0   16  80   0 - 596647 ep_pol 07:14 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
4 S 65532      36012   36002   36012  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36055  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36056  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36057  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36058  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36059  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36066  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36068  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36069  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36070  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36155  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36160  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36161  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36175  0   17  80   0 - 615408 ep_pol 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002   36367  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36012   36002   36368  0   17  80   0 - 615408 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36012   36002  225320  0   17  80   0 - 615408 futex_ 09:04 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S root       36020       1   36020  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/8bb095218306d2f95fb6181c0ecf8e354b80792114b6aac0990452dc4d6ba812/userdata -c 8bb095218306d2f95fb6181c0ecf8e354b80792114b6aac0990452dc4d6ba812 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_neutron-operator-controller-manager-549fb68678-6qbx2_6ef1249c-27f8-4cc0-8134-690b6b8773d1/manager/0.log --log-level info -n k8s_manager_neutron-operator-controller-manager-549fb68678-6qbx2_openstack-operators_6ef1249c-27f8-4cc0-8134-690b6b8773d1_0 -P /run/containers/storage/overlay-containers/8bb095218306d2f95fb6181c0ecf8e354b80792114b6aac0990452dc4d6ba812/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/8bb095218306d2f95fb6181c0ecf8e354b80792114b6aac0990452dc4d6ba812/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/8bb095218306d2f95fb6181c0ecf8e354b80792114b6aac0990452dc4d6ba812/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 8bb095218306d2f95fb6181c0ecf8e354b80792114b6aac0990452dc4d6ba812 -s
1 S root       36021       1   36021  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/48fbec6bb3eceac2d9d66a688ff7dbdacee6092759a8ccc928d3a5e52ab3028d/userdata -c 48fbec6bb3eceac2d9d66a688ff7dbdacee6092759a8ccc928d3a5e52ab3028d --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_keystone-operator-controller-manager-6c9969c6c6-zvpws_21a68e29-471e-428d-9491-ae4b33a01e8a/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_keystone-operator-controller-manager-6c9969c6c6-zvpws_openstack-operators_21a68e29-471e-428d-9491-ae4b33a01e8a_0 -P /run/containers/storage/overlay-containers/48fbec6bb3eceac2d9d66a688ff7dbdacee6092759a8ccc928d3a5e52ab3028d/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/48fbec6bb3eceac2d9d66a688ff7dbdacee6092759a8ccc928d3a5e52ab3028d/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/48fbec6bb3eceac2d9d66a688ff7dbdacee6092759a8ccc928d3a5e52ab3028d/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 48fbec6bb3eceac2d9d66a688ff7dbdacee6092759a8ccc928d3a5e52ab3028d -s
1 S root       36022       1   36022  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/3efcb55a509ac417ec8c6fada6a07436ea5da64e1332f3c9b9c01b0287535dc6/userdata -c 3efcb55a509ac417ec8c6fada6a07436ea5da64e1332f3c9b9c01b0287535dc6 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_infra-operator-controller-manager-7c9978f67-wqn2b_42afdadf-08ba-4196-a815-a4fc0acf2181/manager/0.log --log-level info -n k8s_manager_infra-operator-controller-manager-7c9978f67-wqn2b_openstack-operators_42afdadf-08ba-4196-a815-a4fc0acf2181_0 -P /run/containers/storage/overlay-containers/3efcb55a509ac417ec8c6fada6a07436ea5da64e1332f3c9b9c01b0287535dc6/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/3efcb55a509ac417ec8c6fada6a07436ea5da64e1332f3c9b9c01b0287535dc6/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/3efcb55a509ac417ec8c6fada6a07436ea5da64e1332f3c9b9c01b0287535dc6/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 3efcb55a509ac417ec8c6fada6a07436ea5da64e1332f3c9b9c01b0287535dc6 -s
1 S root       36026       1   36026  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/3bdd9adffe201533f85dcb76185a9db06aac79664ea8bc64ef7576b7aa38cb56/userdata -c 3bdd9adffe201533f85dcb76185a9db06aac79664ea8bc64ef7576b7aa38cb56 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_cinder-operator-controller-manager-8686fd99f7-jlqcq_701c6350-6581-453d-abc4-728bff24f1a5/manager/0.log --log-level info -n k8s_manager_cinder-operator-controller-manager-8686fd99f7-jlqcq_openstack-operators_701c6350-6581-453d-abc4-728bff24f1a5_0 -P /run/containers/storage/overlay-containers/3bdd9adffe201533f85dcb76185a9db06aac79664ea8bc64ef7576b7aa38cb56/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/3bdd9adffe201533f85dcb76185a9db06aac79664ea8bc64ef7576b7aa38cb56/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/3bdd9adffe201533f85dcb76185a9db06aac79664ea8bc64ef7576b7aa38cb56/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 3bdd9adffe201533f85dcb76185a9db06aac79664ea8bc64ef7576b7aa38cb56 -s
4 S 65532      36029   36020   36029  0   17  80   0 - 615689 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   36133  0   17  80   0 - 615689 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   36135  0   17  80   0 - 615689 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   36136  0   17  80   0 - 615689 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   36137  0   17  80   0 - 615689 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   36138  0   17  80   0 - 615689 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36029   36020   36153  0   17  80   0 - 615689 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   36156  0   17  80   0 - 615689 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   36164  0   17  80   0 - 615689 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   36165  0   17  80   0 - 615689 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   36214  0   17  80   0 - 615689 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36029   36020   36215  0   17  80   0 - 615689 ep_pol 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   36306  0   17  80   0 - 615689 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   47488  0   17  80   0 - 615689 futex_ 07:16 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36029   36020   47490  0   17  80   0 - 615689 futex_ 07:16 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   49938  0   17  80   0 - 615689 futex_ 07:16 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36029   36020   49939  0   17  80   0 - 615689 futex_ 07:16 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
4 S 65532      36030   36022   36030  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   36113  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:04 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   36114  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   36115  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   36116  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   36118  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36030   36022   36119  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   36147  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   36148  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:04 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   36149  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   36150  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   36182  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   36183  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   36205  0   17  80   0 - 632639 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   37617  0   17  80   0 - 632639 futex_ 07:14 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   37618  0   17  80   0 - 632639 ep_pol 07:14 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36030   36022   38029  0   17  80   0 - 632639 futex_ 07:14 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
4 S 65532      36031   36021   36031  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36031   36021   36143  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36031   36021   36144  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36031   36021   36145  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36031   36021   36146  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36031   36021   36152  0   11  80   0 - 318162 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36031   36021   36159  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36031   36021   36158  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36031   36021   36162  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36031   36021   36163  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36031   36021   36360  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 65532      36032   36026   36032  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   36123  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   36124  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   36125  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   36126  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   36127  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36032   36026   36134  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36032   36026   36139  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   36151  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36032   36026   36157  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   36201  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   36202  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   36203  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   36237  0   17  80   0 - 615676 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   47228  0   17  80   0 - 615676 futex_ 07:16 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   47229  0   17  80   0 - 615676 ep_pol 07:16 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36032   36026   47230  0   17  80   0 - 615676 futex_ 07:16 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S root       36039       1   36039  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/3eecaf348dd682eb2501f2dbaf49ee77553bcba8399aa6074aa8e7e529249067/userdata -c 3eecaf348dd682eb2501f2dbaf49ee77553bcba8399aa6074aa8e7e529249067 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_barbican-operator-controller-manager-6d6d64fdcf-5d7tm_4686e0fc-c208-4a38-a7ae-33adc6123d0d/manager/0.log --log-level info -n k8s_manager_barbican-operator-controller-manager-6d6d64fdcf-5d7tm_openstack-operators_4686e0fc-c208-4a38-a7ae-33adc6123d0d_0 -P /run/containers/storage/overlay-containers/3eecaf348dd682eb2501f2dbaf49ee77553bcba8399aa6074aa8e7e529249067/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/3eecaf348dd682eb2501f2dbaf49ee77553bcba8399aa6074aa8e7e529249067/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/3eecaf348dd682eb2501f2dbaf49ee77553bcba8399aa6074aa8e7e529249067/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 3eecaf348dd682eb2501f2dbaf49ee77553bcba8399aa6074aa8e7e529249067 -s
1 S root       36040       1   36040  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/d90a59d6065d78809e0e2091bfe39562c87c05ee937de4e0a4141ac0d6223b5f/userdata -c d90a59d6065d78809e0e2091bfe39562c87c05ee937de4e0a4141ac0d6223b5f --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_mariadb-operator-controller-manager-696ff4bcdd-msm2p_9aeddbe0-0fe7-451c-afaf-bd0074505142/manager/0.log --log-level info -n k8s_manager_mariadb-operator-controller-manager-696ff4bcdd-msm2p_openstack-operators_9aeddbe0-0fe7-451c-afaf-bd0074505142_0 -P /run/containers/storage/overlay-containers/d90a59d6065d78809e0e2091bfe39562c87c05ee937de4e0a4141ac0d6223b5f/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/d90a59d6065d78809e0e2091bfe39562c87c05ee937de4e0a4141ac0d6223b5f/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/d90a59d6065d78809e0e2091bfe39562c87c05ee937de4e0a4141ac0d6223b5f/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u d90a59d6065d78809e0e2091bfe39562c87c05ee937de4e0a4141ac0d6223b5f -s
1 S root       36041       1   36041  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/18bfcf2c60443db2649d38972493468f7f1e001a75aa8f2636400a4463edecd0/userdata -c 18bfcf2c60443db2649d38972493468f7f1e001a75aa8f2636400a4463edecd0 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_ironic-operator-controller-manager-59b5fc9845-hv9vj_7ed3bd35-9f00-40d3-9ed0-a111c4131b11/manager/0.log --log-level info -n k8s_manager_ironic-operator-controller-manager-59b5fc9845-hv9vj_openstack-operators_7ed3bd35-9f00-40d3-9ed0-a111c4131b11_0 -P /run/containers/storage/overlay-containers/18bfcf2c60443db2649d38972493468f7f1e001a75aa8f2636400a4463edecd0/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/18bfcf2c60443db2649d38972493468f7f1e001a75aa8f2636400a4463edecd0/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/18bfcf2c60443db2649d38972493468f7f1e001a75aa8f2636400a4463edecd0/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 18bfcf2c60443db2649d38972493468f7f1e001a75aa8f2636400a4463edecd0 -s
1 S root       36042       1   36042  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/72c8d6f575b6e3ec279ab536ba270af237f04f64d08647200bf13c22a0fc147d/userdata -c 72c8d6f575b6e3ec279ab536ba270af237f04f64d08647200bf13c22a0fc147d --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-nthwx_911421e2-0d0c-4e76-b9a7-4cc5cc1ef41d/manager/0.log --log-level info -n k8s_manager_designate-operator-controller-manager-58d86cd59d-nthwx_openstack-operators_911421e2-0d0c-4e76-b9a7-4cc5cc1ef41d_0 -P /run/containers/storage/overlay-containers/72c8d6f575b6e3ec279ab536ba270af237f04f64d08647200bf13c22a0fc147d/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/72c8d6f575b6e3ec279ab536ba270af237f04f64d08647200bf13c22a0fc147d/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/72c8d6f575b6e3ec279ab536ba270af237f04f64d08647200bf13c22a0fc147d/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 72c8d6f575b6e3ec279ab536ba270af237f04f64d08647200bf13c22a0fc147d -s
1 S root       36044       1   36044  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/8f857ed1e9227be61301db3559454dc3f2eb5f73b3e5cf583beb7057b64233ef/userdata -c 8f857ed1e9227be61301db3559454dc3f2eb5f73b3e5cf583beb7057b64233ef --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_heat-operator-controller-manager-5ffbdb7ddf-nf2kl_d812ca0a-067a-4c91-a460-d340ef72d051/manager/0.log --log-level info -n k8s_manager_heat-operator-controller-manager-5ffbdb7ddf-nf2kl_openstack-operators_d812ca0a-067a-4c91-a460-d340ef72d051_0 -P /run/containers/storage/overlay-containers/8f857ed1e9227be61301db3559454dc3f2eb5f73b3e5cf583beb7057b64233ef/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/8f857ed1e9227be61301db3559454dc3f2eb5f73b3e5cf583beb7057b64233ef/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/8f857ed1e9227be61301db3559454dc3f2eb5f73b3e5cf583beb7057b64233ef/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 8f857ed1e9227be61301db3559454dc3f2eb5f73b3e5cf583beb7057b64233ef -s
1 S root       36050       1   36050  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/4e80edf61f94186eb1e5b33db01876e49cc7822899d372fa23029d57ae099b20/userdata -c 4e80edf61f94186eb1e5b33db01876e49cc7822899d372fa23029d57ae099b20 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_glance-operator-controller-manager-d785ddfd5-kln57_e910dcfe-cedc-4a55-ad55-39adb2422c48/manager/0.log --log-level info -n k8s_manager_glance-operator-controller-manager-d785ddfd5-kln57_openstack-operators_e910dcfe-cedc-4a55-ad55-39adb2422c48_0 -P /run/containers/storage/overlay-containers/4e80edf61f94186eb1e5b33db01876e49cc7822899d372fa23029d57ae099b20/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/4e80edf61f94186eb1e5b33db01876e49cc7822899d372fa23029d57ae099b20/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/4e80edf61f94186eb1e5b33db01876e49cc7822899d372fa23029d57ae099b20/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 4e80edf61f94186eb1e5b33db01876e49cc7822899d372fa23029d57ae099b20 -s
4 S 65532      36054   36039   36054  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36054   36039   36243  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36054   36039   36244  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36054   36039   36246  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36054   36039   36255  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36054   36039   36264  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36054   36039   36265  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36054   36039   36314  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36054   36039   36370  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36054   36039   36371  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36054   36039   36446  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36054   36039   36447  0   17  80   0 - 615712 ep_pol 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36054   36039   36465  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36054   36039   36466  0   17  80   0 - 615712 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36054   36039   37956  0   17  80   0 - 615712 futex_ 07:14 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36054   36039   37957  0   17  80   0 - 615712 futex_ 07:14 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36054   36039   37958  0   17  80   0 - 615712 futex_ 07:14 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
4 S 65532      36060   36040   36060  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36060   36040   36274  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36060   36040   36276  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36060   36040   36277  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36060   36040   36278  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36060   36040   36279  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36060   36040   36285  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36060   36040   36321  0   17  80   0 - 631974 ep_pol 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36060   36040   36324  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36060   36040   36364  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36060   36040   36365  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36060   36040   36384  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36060   36040   36385  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36060   36040   36452  0   17  80   0 - 631974 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36060   36040   37900  0   17  80   0 - 631974 futex_ 07:14 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36060   36040   42819  0   17  80   0 - 631974 futex_ 07:15 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36060   36040   42820  0   17  80   0 - 631974 futex_ 07:15 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
4 S 65532      36061   36044   36061  0   17  80   0 - 615536 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   36192  0   17  80   0 - 615536 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   36194  0   17  80   0 - 615536 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   36196  0   17  80   0 - 615536 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   36197  0   17  80   0 - 615536 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   36198  0   17  80   0 - 615536 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   36204  0   17  80   0 - 615536 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36061   36044   36207  0   17  80   0 - 615536 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   36241  0   17  80   0 - 615536 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   36312  0   17  80   0 - 615536 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36061   36044   36330  0   17  80   0 - 615536 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   36343  0   17  80   0 - 615536 ep_pol 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   36345  0   17  80   0 - 615536 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   51410  0   17  80   0 - 615536 futex_ 07:16 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   51411  0   17  80   0 - 615536 futex_ 07:16 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044   51412  0   17  80   0 - 615536 futex_ 07:16 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36061   36044  112381  0   17  80   0 - 615536 futex_ 07:58 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
4 S 65532      36062   36050   36062  0   18  80   0 - 634354 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   36238  0   18  80   0 - 634354 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   36239  0   18  80   0 - 634354 ep_pol 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36062   36050   36247  0   18  80   0 - 634354 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   36256  0   18  80   0 - 634354 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   36261  0   18  80   0 - 634354 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   36262  0   18  80   0 - 634354 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   36263  0   18  80   0 - 634354 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   36320  0   18  80   0 - 634354 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   36351  0   18  80   0 - 634354 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   36355  0   18  80   0 - 634354 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   36356  0   18  80   0 - 634354 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36062   36050   36357  0   18  80   0 - 634354 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   37935  0   18  80   0 - 634354 futex_ 07:14 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   37936  0   18  80   0 - 634354 futex_ 07:14 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36062   36050   42264  0   18  80   0 - 634354 futex_ 07:15 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   42265  0   18  80   0 - 634354 futex_ 07:15 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36062   36050   49428  0   18  80   0 - 634354 futex_ 07:16 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
4 S 65532      36072   36042   36072  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36248  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:03 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36249  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36251  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36257  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36258  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36291  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36346  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36347  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36348  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36380  0   17  80   0 - 632169 ep_pol 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36381  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36549  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36550  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36551  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36552  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36072   36042   36554  0   17  80   0 - 632169 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
4 S 65532      36075   36041   36075  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36170  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:03 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36171  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36172  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36173  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36174  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36176  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36177  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36180  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36186  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36254  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36290  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36310  0   14  80   0 - 576699 ep_pol 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36075   36041   36316  0   14  80   0 - 576699 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S root       36076       1   36076  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/dd7e01494b63a49be075273cfbb57e40f41979037d03615dbb394870ae5a8e86/userdata -c dd7e01494b63a49be075273cfbb57e40f41979037d03615dbb394870ae5a8e86 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_octavia-operator-controller-manager-b4444585c-9qkqx_b1578049-8763-4c26-b149-8497d94da92e/manager/0.log --log-level info -n k8s_manager_octavia-operator-controller-manager-b4444585c-9qkqx_openstack-operators_b1578049-8763-4c26-b149-8497d94da92e_0 -P /run/containers/storage/overlay-containers/dd7e01494b63a49be075273cfbb57e40f41979037d03615dbb394870ae5a8e86/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/dd7e01494b63a49be075273cfbb57e40f41979037d03615dbb394870ae5a8e86/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/dd7e01494b63a49be075273cfbb57e40f41979037d03615dbb394870ae5a8e86/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u dd7e01494b63a49be075273cfbb57e40f41979037d03615dbb394870ae5a8e86 -s
4 S 65532      36084   36076   36084  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36233  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36235  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36236  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36250  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36284  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36300  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36361  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36362  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36366  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36396  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36474  0   14  80   0 - 560456 ep_pol 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36475  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36084   36076   36476  0   14  80   0 - 560456 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S root       36095       1   36095  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/08d0853075ca3019271df143af6aeee57c2b0d9edab417777317e80c3acb6ff5/userdata -c 08d0853075ca3019271df143af6aeee57c2b0d9edab417777317e80c3acb6ff5 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-7dv7x_1a5d562f-6fe0-4bc9-a39a-7f7d8e9c1b63/manager/0.log --log-level info -n k8s_manager_swift-operator-controller-manager-76d5577b-7dv7x_openstack-operators_1a5d562f-6fe0-4bc9-a39a-7f7d8e9c1b63_0 -P /run/containers/storage/overlay-containers/08d0853075ca3019271df143af6aeee57c2b0d9edab417777317e80c3acb6ff5/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/08d0853075ca3019271df143af6aeee57c2b0d9edab417777317e80c3acb6ff5/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/08d0853075ca3019271df143af6aeee57c2b0d9edab417777317e80c3acb6ff5/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 08d0853075ca3019271df143af6aeee57c2b0d9edab417777317e80c3acb6ff5 -s
1 S root       36096       1   36096  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e4b3233e7cfd5ace801f9a818d6b2241823119c76f124ac0e3b46ffba6a67387/userdata -c e4b3233e7cfd5ace801f9a818d6b2241823119c76f124ac0e3b46ffba6a67387 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_manila-operator-controller-manager-66fdd975d9-47p2c_6c644f74-4957-40ad-8286-b172b802a323/manager/0.log --log-level info -n k8s_manager_manila-operator-controller-manager-66fdd975d9-47p2c_openstack-operators_6c644f74-4957-40ad-8286-b172b802a323_0 -P /run/containers/storage/overlay-containers/e4b3233e7cfd5ace801f9a818d6b2241823119c76f124ac0e3b46ffba6a67387/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e4b3233e7cfd5ace801f9a818d6b2241823119c76f124ac0e3b46ffba6a67387/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e4b3233e7cfd5ace801f9a818d6b2241823119c76f124ac0e3b46ffba6a67387/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e4b3233e7cfd5ace801f9a818d6b2241823119c76f124ac0e3b46ffba6a67387 -s
1 S root       36099       1   36099  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/ff42589d41455b1f1b86f157e377a883f75ab80aece2b392d8deed458fa22e34/userdata -c ff42589d41455b1f1b86f157e377a883f75ab80aece2b392d8deed458fa22e34 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_nova-operator-controller-manager-5b45478b88-cj2zz_5fa60d72-b0e4-4a27-b2af-d228cd9db5da/manager/0.log --log-level info -n k8s_manager_nova-operator-controller-manager-5b45478b88-cj2zz_openstack-operators_5fa60d72-b0e4-4a27-b2af-d228cd9db5da_0 -P /run/containers/storage/overlay-containers/ff42589d41455b1f1b86f157e377a883f75ab80aece2b392d8deed458fa22e34/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/ff42589d41455b1f1b86f157e377a883f75ab80aece2b392d8deed458fa22e34/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/ff42589d41455b1f1b86f157e377a883f75ab80aece2b392d8deed458fa22e34/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u ff42589d41455b1f1b86f157e377a883f75ab80aece2b392d8deed458fa22e34 -s
1 S root       36101       1   36101  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/6daad7a4925a17b85f45e6448589fa6a640f48349b5ecbede95de400bfc14440/userdata -c 6daad7a4925a17b85f45e6448589fa6a640f48349b5ecbede95de400bfc14440 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_telemetry-operator-controller-manager-5ffb97cddf-vf8h7_368ec6b5-ba12-467e-ab4c-d46a83c31483/manager/0.log --log-level info -n k8s_manager_telemetry-operator-controller-manager-5ffb97cddf-vf8h7_openstack-operators_368ec6b5-ba12-467e-ab4c-d46a83c31483_0 -P /run/containers/storage/overlay-containers/6daad7a4925a17b85f45e6448589fa6a640f48349b5ecbede95de400bfc14440/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/6daad7a4925a17b85f45e6448589fa6a640f48349b5ecbede95de400bfc14440/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/6daad7a4925a17b85f45e6448589fa6a640f48349b5ecbede95de400bfc14440/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 6daad7a4925a17b85f45e6448589fa6a640f48349b5ecbede95de400bfc14440 -s
4 S 65532      36103   36099   36103  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36103   36099   36304  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36103   36099   36305  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36103   36099   36307  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36103   36099   36308  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36103   36099   36309  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36103   36099   36311  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36103   36099   36313  0   18  80   0 - 634516 ep_pol 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36103   36099   36328  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36103   36099   36389  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36103   36099   36424  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36103   36099   36472  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36103   36099   36473  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36103   36099   36518  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36103   36099   36519  0   18  80   0 - 634516 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36103   36099   58848  0   18  80   0 - 634516 futex_ 07:17 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36103   36099   58849  0   18  80   0 - 634516 futex_ 07:17 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36103   36099   69275  0   18  80   0 - 634516 futex_ 07:19 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
4 S 65532      36105   36101   36105  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36105   36101   36191  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36105   36101   36193  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36105   36101   36195  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36105   36101   36199  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36105   36101   36200  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36105   36101   36209  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36105   36101   36282  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36105   36101   36283  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36105   36101   36286  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36105   36101   36339  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36105   36101   36340  0   17  80   0 - 632822 ep_pol 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36105   36101   36342  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36105   36101   36378  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36105   36101   36379  0   17  80   0 - 632822 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36105   36101   45386  0   17  80   0 - 632822 futex_ 07:15 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36105   36101   54264  0   17  80   0 - 632822 futex_ 07:17 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
4 S 1000660+   36109   36095   36109  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36109   36095   36294  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:03 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   36109   36095   36295  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36109   36095   36296  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36109   36095   36297  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36109   36095   36299  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36109   36095   36317  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36109   36095   36382  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36109   36095   36387  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   36109   36095   36392  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36109   36095   36393  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36109   36095   36454  0   18  80   0 - 650720 ep_pol 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36109   36095   36458  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   36109   36095   36459  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   36109   36095   36461  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36109   36095   36462  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 1000660+   36109   36095   36463  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36109   36095   36477  0   18  80   0 - 650720 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
4 S 65532      36110   36096   36110  0   13  80   0 - 541595 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36110   36096   36229  0   13  80   0 - 541595 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36110   36096   36230  0   13  80   0 - 541595 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36110   36096   36231  0   13  80   0 - 541595 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36110   36096   36232  0   13  80   0 - 541595 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36110   36096   36259  0   13  80   0 - 541595 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36110   36096   36260  0   13  80   0 - 541595 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36110   36096   36268  0   13  80   0 - 541595 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36110   36096   36269  0   13  80   0 - 541595 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36110   36096   36338  0   13  80   0 - 541595 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36110   36096   36359  0   13  80   0 - 541595 ep_pol 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36110   36096   36373  0   13  80   0 - 541595 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36110   36096   36391  0   13  80   0 - 541595 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S root       36130       1   36130  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/9c2755e06c490413c54cbe393d235b8b65a078507ec706627c762975aac20a4a/userdata -c 9c2755e06c490413c54cbe393d235b8b65a078507ec706627c762975aac20a4a --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_watcher-operator-controller-manager-5595cf6c95-4hcbj_9ab8a530-fefb-477b-85f7-6f716f684292/manager/0.log --log-level info -n k8s_manager_watcher-operator-controller-manager-5595cf6c95-4hcbj_openstack-operators_9ab8a530-fefb-477b-85f7-6f716f684292_0 -P /run/containers/storage/overlay-containers/9c2755e06c490413c54cbe393d235b8b65a078507ec706627c762975aac20a4a/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/9c2755e06c490413c54cbe393d235b8b65a078507ec706627c762975aac20a4a/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/9c2755e06c490413c54cbe393d235b8b65a078507ec706627c762975aac20a4a/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 9c2755e06c490413c54cbe393d235b8b65a078507ec706627c762975aac20a4a -s
4 S 65532      36140   36130   36140  0   13  80   0 - 541619 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36140   36130   36331  0   13  80   0 - 541619 futex_ 07:13 ?       00:00:03 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36140   36130   36332  0   13  80   0 - 541619 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36140   36130   36333  0   13  80   0 - 541619 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36140   36130   36334  0   13  80   0 - 541619 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36140   36130   36337  0   13  80   0 - 541619 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36140   36130   36352  0   13  80   0 - 541619 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36140   36130   36353  0   13  80   0 - 541619 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36140   36130   36394  0   13  80   0 - 541619 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36140   36130   36395  0   13  80   0 - 541619 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36140   36130   36397  0   13  80   0 - 541619 ep_pol 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36140   36130   36504  0   13  80   0 - 541619 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36140   36130   36525  0   13  80   0 - 541619 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S root       36178       1   36178  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/826bbc726955158f51ef24bfbe18790e0162951d6b56d23b4f0454bb61bb30fa/userdata -c 826bbc726955158f51ef24bfbe18790e0162951d6b56d23b4f0454bb61bb30fa --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_ovn-operator-controller-manager-855d7949fc-gnvp7_2514440b-b998-4078-834e-642c3bcae80f/manager/0.log --log-level info -n k8s_manager_ovn-operator-controller-manager-855d7949fc-gnvp7_openstack-operators_2514440b-b998-4078-834e-642c3bcae80f_0 -P /run/containers/storage/overlay-containers/826bbc726955158f51ef24bfbe18790e0162951d6b56d23b4f0454bb61bb30fa/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/826bbc726955158f51ef24bfbe18790e0162951d6b56d23b4f0454bb61bb30fa/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/826bbc726955158f51ef24bfbe18790e0162951d6b56d23b4f0454bb61bb30fa/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 826bbc726955158f51ef24bfbe18790e0162951d6b56d23b4f0454bb61bb30fa -s
4 S 65532      36181   36178   36181  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36181   36178   36409  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36181   36178   36414  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36181   36178   36416  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36181   36178   36417  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36181   36178   36418  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36181   36178   36419  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36181   36178   36445  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36181   36178   36467  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36181   36178   36468  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36181   36178   36526  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36181   36178   36533  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36181   36178   36539  0   17  80   0 - 631852 ep_pol 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36181   36178   36540  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36181   36178   36543  0   17  80   0 - 631852 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36181   36178   38164  0   17  80   0 - 631852 futex_ 07:14 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      36181   36178   38165  0   17  80   0 - 631852 futex_ 07:14 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S root       36217       1   36217  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/5d1f03db10fb5ecdae545d2a8e8dcf3a105b033cc5d3c27e2605226a6cd97ebe/userdata -c 5d1f03db10fb5ecdae545d2a8e8dcf3a105b033cc5d3c27e2605226a6cd97ebe --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_placement-operator-controller-manager-ccbfcb8c-tnrxp_b842fd5e-6125-40cf-b729-947e54309c87/manager/0.log --log-level info -n k8s_manager_placement-operator-controller-manager-ccbfcb8c-tnrxp_openstack-operators_b842fd5e-6125-40cf-b729-947e54309c87_0 -P /run/containers/storage/overlay-containers/5d1f03db10fb5ecdae545d2a8e8dcf3a105b033cc5d3c27e2605226a6cd97ebe/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/5d1f03db10fb5ecdae545d2a8e8dcf3a105b033cc5d3c27e2605226a6cd97ebe/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/5d1f03db10fb5ecdae545d2a8e8dcf3a105b033cc5d3c27e2605226a6cd97ebe/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 5d1f03db10fb5ecdae545d2a8e8dcf3a105b033cc5d3c27e2605226a6cd97ebe -s
1 S root       36221       1   36221  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/7d29cb89fef39203f3be21d499b31bda5d39afcc2a1faac3e571798c5b296740/userdata -c 7d29cb89fef39203f3be21d499b31bda5d39afcc2a1faac3e571798c5b296740 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_neutron-operator-controller-manager-549fb68678-6qbx2_6ef1249c-27f8-4cc0-8134-690b6b8773d1/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_neutron-operator-controller-manager-549fb68678-6qbx2_openstack-operators_6ef1249c-27f8-4cc0-8134-690b6b8773d1_0 -P /run/containers/storage/overlay-containers/7d29cb89fef39203f3be21d499b31bda5d39afcc2a1faac3e571798c5b296740/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/7d29cb89fef39203f3be21d499b31bda5d39afcc2a1faac3e571798c5b296740/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/7d29cb89fef39203f3be21d499b31bda5d39afcc2a1faac3e571798c5b296740/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 7d29cb89fef39203f3be21d499b31bda5d39afcc2a1faac3e571798c5b296740 -s
4 S 65532      36223   36217   36223  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36374  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36375  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36376  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36377  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36383  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36398  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36399  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36404  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36498  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36502  0   17  80   0 - 615521 ep_pol 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36512  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36523  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36653  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36654  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36655  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      36223   36217   36656  0   17  80   0 - 615521 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
4 S 65532      36245   36221   36245  0   10  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36245   36221   36433  0   10  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36245   36221   36434  0   10  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36245   36221   36435  0   10  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36245   36221   36436  0   10  80   0 - 318098 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36245   36221   36520  0   10  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36245   36221   36521  0   10  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36245   36221   36564  0   10  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36245   36221   44711  0   10  80   0 - 318098 futex_ 07:15 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36245   36221   44712  0   10  80   0 - 318098 futex_ 07:15 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S root       36271       1   36271  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/fc0668b1a2f61c50e11aab7f321d85e4a9157f89a7247aad244c57c647ef526d/userdata -c fc0668b1a2f61c50e11aab7f321d85e4a9157f89a7247aad244c57c647ef526d --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-bcntr_52bc186b-754c-4fc1-982d-435c345718a7/operator/0.log --log-level info -n k8s_operator_rabbitmq-cluster-operator-manager-5f97d8c699-bcntr_openstack-operators_52bc186b-754c-4fc1-982d-435c345718a7_0 -P /run/containers/storage/overlay-containers/fc0668b1a2f61c50e11aab7f321d85e4a9157f89a7247aad244c57c647ef526d/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/fc0668b1a2f61c50e11aab7f321d85e4a9157f89a7247aad244c57c647ef526d/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/fc0668b1a2f61c50e11aab7f321d85e4a9157f89a7247aad244c57c647ef526d/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u fc0668b1a2f61c50e11aab7f321d85e4a9157f89a7247aad244c57c647ef526d -s
4 S 1000660+   36275   36271   36275  0   17  80   0 - 615305 futex_ 07:13 ?       00:00:01 /manager
1 S 1000660+   36275   36271   36410  0   17  80   0 - 615305 futex_ 07:13 ?       00:00:01 /manager
1 S 1000660+   36275   36271   36411  0   17  80   0 - 615305 futex_ 07:13 ?       00:00:00 /manager
1 S 1000660+   36275   36271   36412  0   17  80   0 - 615305 futex_ 07:13 ?       00:00:00 /manager
1 S 1000660+   36275   36271   36413  0   17  80   0 - 615305 futex_ 07:13 ?       00:00:00 /manager
1 S 1000660+   36275   36271   36428  0   17  80   0 - 615305 futex_ 07:13 ?       00:00:00 /manager
1 S 1000660+   36275   36271   36451  0   17  80   0 - 615305 futex_ 07:13 ?       00:00:00 /manager
1 S 1000660+   36275   36271   36460  0   17  80   0 - 615305 futex_ 07:13 ?       00:00:00 /manager
1 S 1000660+   36275   36271   36516  0   17  80   0 - 615305 futex_ 07:13 ?       00:00:00 /manager
1 S 1000660+   36275   36271   36522  0   17  80   0 - 615305 futex_ 07:13 ?       00:00:00 /manager
1 S 1000660+   36275   36271   37619  0   17  80   0 - 615305 futex_ 07:14 ?       00:00:00 /manager
1 S 1000660+   36275   36271   37620  0   17  80   0 - 615305 futex_ 07:14 ?       00:00:00 /manager
1 S 1000660+   36275   36271   42151  0   17  80   0 - 615305 futex_ 07:15 ?       00:00:00 /manager
1 S 1000660+   36275   36271   42152  0   17  80   0 - 615305 futex_ 07:15 ?       00:00:00 /manager
1 S 1000660+   36275   36271   42153  0   17  80   0 - 615305 futex_ 07:15 ?       00:00:00 /manager
1 S 1000660+   36275   36271   42154  0   17  80   0 - 615305 futex_ 07:15 ?       00:00:00 /manager
1 S 1000660+   36275   36271  158325  0   17  80   0 - 615305 ep_pol 08:39 ?       00:00:00 /manager
1 S root       36287       1   36287  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e569c7ec9baef5f47c9c62bccddc74fa7ef0cb03ddafc563607359a203481fe8/userdata -c e569c7ec9baef5f47c9c62bccddc74fa7ef0cb03ddafc563607359a203481fe8 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_cinder-operator-controller-manager-8686fd99f7-jlqcq_701c6350-6581-453d-abc4-728bff24f1a5/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_cinder-operator-controller-manager-8686fd99f7-jlqcq_openstack-operators_701c6350-6581-453d-abc4-728bff24f1a5_0 -P /run/containers/storage/overlay-containers/e569c7ec9baef5f47c9c62bccddc74fa7ef0cb03ddafc563607359a203481fe8/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e569c7ec9baef5f47c9c62bccddc74fa7ef0cb03ddafc563607359a203481fe8/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e569c7ec9baef5f47c9c62bccddc74fa7ef0cb03ddafc563607359a203481fe8/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e569c7ec9baef5f47c9c62bccddc74fa7ef0cb03ddafc563607359a203481fe8 -s
4 S 65532      36292   36287   36292  0   11  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36292   36287   36421  0   11  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36292   36287   36422  0   11  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36292   36287   36423  0   11  80   0 - 318226 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36292   36287   36425  0   11  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36292   36287   36469  0   11  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36292   36287   36470  0   11  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36292   36287   36499  0   11  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36292   36287   36501  0   11  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36292   36287   36577  0   11  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36292   36287   89828  0   11  80   0 - 318226 futex_ 07:36 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S root       36298       1   36298  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/d8291b2c62e36b2366652d87b49c99ecb25bcbeb58df7b8da0c1cc1596fbd939/userdata -c d8291b2c62e36b2366652d87b49c99ecb25bcbeb58df7b8da0c1cc1596fbd939 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_horizon-operator-controller-manager-586b66cf4f-qxv5t_84c853db-41b6-4013-b6a6-b9f6c3fa74e3/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_horizon-operator-controller-manager-586b66cf4f-qxv5t_openstack-operators_84c853db-41b6-4013-b6a6-b9f6c3fa74e3_0 -P /run/containers/storage/overlay-containers/d8291b2c62e36b2366652d87b49c99ecb25bcbeb58df7b8da0c1cc1596fbd939/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/d8291b2c62e36b2366652d87b49c99ecb25bcbeb58df7b8da0c1cc1596fbd939/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/d8291b2c62e36b2366652d87b49c99ecb25bcbeb58df7b8da0c1cc1596fbd939/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u d8291b2c62e36b2366652d87b49c99ecb25bcbeb58df7b8da0c1cc1596fbd939 -s
4 S 65532      36315   36298   36315  0   10  80   0 - 318034 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36315   36298   36437  0   10  80   0 - 318034 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36315   36298   36438  0   10  80   0 - 318034 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36315   36298   36439  0   10  80   0 - 318034 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36315   36298   36440  0   10  80   0 - 318034 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36315   36298   36443  0   10  80   0 - 318034 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36315   36298   36471  0   10  80   0 - 318034 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36315   36298   36590  0   10  80   0 - 318034 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36315   36298   45087  0   10  80   0 - 318034 futex_ 07:15 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36315   36298   45088  0   10  80   0 - 318034 futex_ 07:15 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S root       36322       1   36322  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/5d885d2b3aafd75766e564536a142493039d291045a3e2153001b32e8b40f200/userdata -c 5d885d2b3aafd75766e564536a142493039d291045a3e2153001b32e8b40f200 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_test-operator-controller-manager-6bb6dcddc-hmtgk_d6c2b423-4398-416f-bb17-70a8eb814964/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_test-operator-controller-manager-6bb6dcddc-hmtgk_openstack-operators_d6c2b423-4398-416f-bb17-70a8eb814964_0 -P /run/containers/storage/overlay-containers/5d885d2b3aafd75766e564536a142493039d291045a3e2153001b32e8b40f200/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/5d885d2b3aafd75766e564536a142493039d291045a3e2153001b32e8b40f200/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/5d885d2b3aafd75766e564536a142493039d291045a3e2153001b32e8b40f200/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 5d885d2b3aafd75766e564536a142493039d291045a3e2153001b32e8b40f200 -s
4 S 65532      36325   36322   36325  0   12  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36325   36322   36455  0   12  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36325   36322   36456  0   12  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36325   36322   36457  0   12  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36325   36322   36464  0   12  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36325   36322   36478  0   12  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36325   36322   36505  0   12  80   0 - 318226 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36325   36322   36578  0   12  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36325   36322   36579  0   12  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36325   36322   36580  0   12  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36325   36322   36581  0   12  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36325   36322   36683  0   12  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S root       36335       1   36335  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/1a07ec8a18e8878e641f413459b52286b63e46037f560ab9fef8f43f2a57bc4d/userdata -c 1a07ec8a18e8878e641f413459b52286b63e46037f560ab9fef8f43f2a57bc4d --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_ironic-operator-controller-manager-59b5fc9845-hv9vj_7ed3bd35-9f00-40d3-9ed0-a111c4131b11/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_ironic-operator-controller-manager-59b5fc9845-hv9vj_openstack-operators_7ed3bd35-9f00-40d3-9ed0-a111c4131b11_0 -P /run/containers/storage/overlay-containers/1a07ec8a18e8878e641f413459b52286b63e46037f560ab9fef8f43f2a57bc4d/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/1a07ec8a18e8878e641f413459b52286b63e46037f560ab9fef8f43f2a57bc4d/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/1a07ec8a18e8878e641f413459b52286b63e46037f560ab9fef8f43f2a57bc4d/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 1a07ec8a18e8878e641f413459b52286b63e46037f560ab9fef8f43f2a57bc4d -s
4 S 65532      36344   36335   36344  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36344   36335   36529  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36344   36335   36530  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36344   36335   36531  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36344   36335   36532  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36344   36335   36534  0   11  80   0 - 318162 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36344   36335   36535  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36344   36335   36536  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36344   36335   36537  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36344   36335   36538  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36344   36335   36680  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S root       36349       1   36349  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/5d9d22e44a4c391c9cd74ba0aa5f30672b77822953523df58c1de933b6ca0215/userdata -c 5d9d22e44a4c391c9cd74ba0aa5f30672b77822953523df58c1de933b6ca0215 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_infra-operator-controller-manager-7c9978f67-wqn2b_42afdadf-08ba-4196-a815-a4fc0acf2181/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_infra-operator-controller-manager-7c9978f67-wqn2b_openstack-operators_42afdadf-08ba-4196-a815-a4fc0acf2181_0 -P /run/containers/storage/overlay-containers/5d9d22e44a4c391c9cd74ba0aa5f30672b77822953523df58c1de933b6ca0215/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/5d9d22e44a4c391c9cd74ba0aa5f30672b77822953523df58c1de933b6ca0215/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/5d9d22e44a4c391c9cd74ba0aa5f30672b77822953523df58c1de933b6ca0215/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 5d9d22e44a4c391c9cd74ba0aa5f30672b77822953523df58c1de933b6ca0215 -s
4 S 65532      36354   36349   36354  0   13  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36354   36349   36584  0   13  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36354   36349   36585  0   13  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36354   36349   36586  0   13  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36354   36349   36587  0   13  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36354   36349   36588  0   13  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36354   36349   36589  0   13  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36354   36349   36591  0   13  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36354   36349   36598  0   13  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36354   36349   36599  0   13  80   0 - 318354 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36354   36349   36600  0   13  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36354   36349   36601  0   13  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36354   36349   36689  0   13  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S root       36479       1   36479  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/5e9044dfec0868db729fc1951674f4ab7d4402bd3b2ecec2605d42a0bb60340e/userdata -c 5e9044dfec0868db729fc1951674f4ab7d4402bd3b2ecec2605d42a0bb60340e --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-nthwx_911421e2-0d0c-4e76-b9a7-4cc5cc1ef41d/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_designate-operator-controller-manager-58d86cd59d-nthwx_openstack-operators_911421e2-0d0c-4e76-b9a7-4cc5cc1ef41d_0 -P /run/containers/storage/overlay-containers/5e9044dfec0868db729fc1951674f4ab7d4402bd3b2ecec2605d42a0bb60340e/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/5e9044dfec0868db729fc1951674f4ab7d4402bd3b2ecec2605d42a0bb60340e/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/5e9044dfec0868db729fc1951674f4ab7d4402bd3b2ecec2605d42a0bb60340e/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 5e9044dfec0868db729fc1951674f4ab7d4402bd3b2ecec2605d42a0bb60340e -s
1 S root       36480       1   36480  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/eaaf2c37a52645903e5243a07522d80da57692ca002cfc54e26f7a6f175dbff6/userdata -c eaaf2c37a52645903e5243a07522d80da57692ca002cfc54e26f7a6f175dbff6 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_nova-operator-controller-manager-5b45478b88-cj2zz_5fa60d72-b0e4-4a27-b2af-d228cd9db5da/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_nova-operator-controller-manager-5b45478b88-cj2zz_openstack-operators_5fa60d72-b0e4-4a27-b2af-d228cd9db5da_0 -P /run/containers/storage/overlay-containers/eaaf2c37a52645903e5243a07522d80da57692ca002cfc54e26f7a6f175dbff6/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/eaaf2c37a52645903e5243a07522d80da57692ca002cfc54e26f7a6f175dbff6/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/eaaf2c37a52645903e5243a07522d80da57692ca002cfc54e26f7a6f175dbff6/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u eaaf2c37a52645903e5243a07522d80da57692ca002cfc54e26f7a6f175dbff6 -s
1 S root       36481       1   36481  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/aa2319d2c14e46dd39ede9e08dad96bb4ec7d30acbe9eb054ee2700125d56fc4/userdata -c aa2319d2c14e46dd39ede9e08dad96bb4ec7d30acbe9eb054ee2700125d56fc4 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_heat-operator-controller-manager-5ffbdb7ddf-nf2kl_d812ca0a-067a-4c91-a460-d340ef72d051/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_heat-operator-controller-manager-5ffbdb7ddf-nf2kl_openstack-operators_d812ca0a-067a-4c91-a460-d340ef72d051_0 -P /run/containers/storage/overlay-containers/aa2319d2c14e46dd39ede9e08dad96bb4ec7d30acbe9eb054ee2700125d56fc4/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/aa2319d2c14e46dd39ede9e08dad96bb4ec7d30acbe9eb054ee2700125d56fc4/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/aa2319d2c14e46dd39ede9e08dad96bb4ec7d30acbe9eb054ee2700125d56fc4/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u aa2319d2c14e46dd39ede9e08dad96bb4ec7d30acbe9eb054ee2700125d56fc4 -s
1 S root       36483       1   36483  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/e6f9a820b5fb49387fd3d2ce1fcfcb406e7412b82024f277e75d4a71d5514565/userdata -c e6f9a820b5fb49387fd3d2ce1fcfcb406e7412b82024f277e75d4a71d5514565 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_barbican-operator-controller-manager-6d6d64fdcf-5d7tm_4686e0fc-c208-4a38-a7ae-33adc6123d0d/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_barbican-operator-controller-manager-6d6d64fdcf-5d7tm_openstack-operators_4686e0fc-c208-4a38-a7ae-33adc6123d0d_0 -P /run/containers/storage/overlay-containers/e6f9a820b5fb49387fd3d2ce1fcfcb406e7412b82024f277e75d4a71d5514565/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/e6f9a820b5fb49387fd3d2ce1fcfcb406e7412b82024f277e75d4a71d5514565/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/e6f9a820b5fb49387fd3d2ce1fcfcb406e7412b82024f277e75d4a71d5514565/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u e6f9a820b5fb49387fd3d2ce1fcfcb406e7412b82024f277e75d4a71d5514565 -s
1 S root       36484       1   36484  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/5b3745ba6465713bb298f33b25e725d53c45e2bda03939230f864cef6b945217/userdata -c 5b3745ba6465713bb298f33b25e725d53c45e2bda03939230f864cef6b945217 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_octavia-operator-controller-manager-b4444585c-9qkqx_b1578049-8763-4c26-b149-8497d94da92e/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_octavia-operator-controller-manager-b4444585c-9qkqx_openstack-operators_b1578049-8763-4c26-b149-8497d94da92e_0 -P /run/containers/storage/overlay-containers/5b3745ba6465713bb298f33b25e725d53c45e2bda03939230f864cef6b945217/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/5b3745ba6465713bb298f33b25e725d53c45e2bda03939230f864cef6b945217/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/5b3745ba6465713bb298f33b25e725d53c45e2bda03939230f864cef6b945217/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 5b3745ba6465713bb298f33b25e725d53c45e2bda03939230f864cef6b945217 -s
1 S root       36488       1   36488  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/8a57d89cd16753d00a9ad4599dee793fc8982636937d6560cb34d36d282c1871/userdata -c 8a57d89cd16753d00a9ad4599dee793fc8982636937d6560cb34d36d282c1871 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_mariadb-operator-controller-manager-696ff4bcdd-msm2p_9aeddbe0-0fe7-451c-afaf-bd0074505142/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_mariadb-operator-controller-manager-696ff4bcdd-msm2p_openstack-operators_9aeddbe0-0fe7-451c-afaf-bd0074505142_0 -P /run/containers/storage/overlay-containers/8a57d89cd16753d00a9ad4599dee793fc8982636937d6560cb34d36d282c1871/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/8a57d89cd16753d00a9ad4599dee793fc8982636937d6560cb34d36d282c1871/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/8a57d89cd16753d00a9ad4599dee793fc8982636937d6560cb34d36d282c1871/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 8a57d89cd16753d00a9ad4599dee793fc8982636937d6560cb34d36d282c1871 -s
1 S root       36490       1   36490  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/daff92d4642b3deee4a4940d041f66cc1a5cc8b81d380069a97132f37f8a3175/userdata -c daff92d4642b3deee4a4940d041f66cc1a5cc8b81d380069a97132f37f8a3175 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_manila-operator-controller-manager-66fdd975d9-47p2c_6c644f74-4957-40ad-8286-b172b802a323/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_manila-operator-controller-manager-66fdd975d9-47p2c_openstack-operators_6c644f74-4957-40ad-8286-b172b802a323_0 -P /run/containers/storage/overlay-containers/daff92d4642b3deee4a4940d041f66cc1a5cc8b81d380069a97132f37f8a3175/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/daff92d4642b3deee4a4940d041f66cc1a5cc8b81d380069a97132f37f8a3175/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/daff92d4642b3deee4a4940d041f66cc1a5cc8b81d380069a97132f37f8a3175/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u daff92d4642b3deee4a4940d041f66cc1a5cc8b81d380069a97132f37f8a3175 -s
1 S root       36492       1   36492  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/97f6f433254cbd8b17c6a7d32c7bab1df852c9744c6a9fa27e897b085438350f/userdata -c 97f6f433254cbd8b17c6a7d32c7bab1df852c9744c6a9fa27e897b085438350f --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_glance-operator-controller-manager-d785ddfd5-kln57_e910dcfe-cedc-4a55-ad55-39adb2422c48/kube-rbac-proxy/0.log --log-level info -n k8s_kube-rbac-proxy_glance-operator-controller-manager-d785ddfd5-kln57_openstack-operators_e910dcfe-cedc-4a55-ad55-39adb2422c48_0 -P /run/containers/storage/overlay-containers/97f6f433254cbd8b17c6a7d32c7bab1df852c9744c6a9fa27e897b085438350f/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/97f6f433254cbd8b17c6a7d32c7bab1df852c9744c6a9fa27e897b085438350f/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/97f6f433254cbd8b17c6a7d32c7bab1df852c9744c6a9fa27e897b085438350f/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 97f6f433254cbd8b17c6a7d32c7bab1df852c9744c6a9fa27e897b085438350f -s
4 S 65532      36493   36481   36493  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36493   36481   36616  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36493   36481   36618  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36493   36481   36619  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36493   36481   36642  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36493   36481   36644  0   11  80   0 - 318162 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36493   36481   36658  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36493   36481   36659  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36493   36481   36686  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36493   36481   36687  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36493   36481  131745  0   11  80   0 - 318162 futex_ 08:16 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 65532      36496   36484   36496  0   12  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36496   36484   36594  0   12  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36496   36484   36595  0   12  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36496   36484   36596  0   12  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36496   36484   36597  0   12  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36496   36484   36602  0   12  80   0 - 318354 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36496   36484   36610  0   12  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36496   36484   36611  0   12  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36496   36484   36612  0   12  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36496   36484   36651  0   12  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36496   36484   36652  0   12  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36496   36484   36688  0   12  80   0 - 318354 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 65532      36497   36488   36497  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36497   36488   36631  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36497   36488   36633  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36497   36488   36635  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36497   36488   36639  0   10  80   0 - 318162 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36497   36488   36650  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36497   36488   36664  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36497   36488   36665  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36497   36488   36666  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36497   36488   36681  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 65532      36500   36479   36500  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36500   36479   36613  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36500   36479   36614  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36500   36479   36615  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36500   36479   36617  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36500   36479   36660  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36500   36479   36675  0   10  80   0 - 318162 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36500   36479   36684  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36500   36479   36685  0   10  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36500   36479   45120  0   10  80   0 - 318162 futex_ 07:15 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 65532      36503   36480   36503  0   11  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36503   36480   36626  0   11  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36503   36480   36627  0   11  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36503   36480   36628  0   11  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36503   36480   36629  0   11  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36503   36480   36667  0   11  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36503   36480   36676  0   11  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36503   36480   36677  0   11  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36503   36480   36682  0   11  80   0 - 318098 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36503   36480   75378  0   11  80   0 - 318098 futex_ 07:21 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36503   36480  215555  0   11  80   0 - 318098 ep_pol 09:01 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 65532      36506   36490   36506  0   10  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36506   36490   36632  0   10  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36506   36490   36634  0   10  80   0 - 318226 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36506   36490   36636  0   10  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36506   36490   36637  0   10  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36506   36490   36638  0   10  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36506   36490   36661  0   10  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36506   36490   36662  0   10  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36506   36490   36663  0   10  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36506   36490   36693  0   10  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 65532      36510   36492   36510  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36510   36492   36645  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36510   36492   36646  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36510   36492   36647  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36510   36492   36648  0   11  80   0 - 318162 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36510   36492   36668  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36510   36492   36669  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36510   36492   36670  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36510   36492   36678  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36510   36492   36679  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36510   36492   36692  0   11  80   0 - 318162 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
4 S 65532      36513   36483   36513  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36513   36483   36623  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36513   36483   36624  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36513   36483   36625  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36513   36483   36630  0   13  80   0 - 318226 ep_pol 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36513   36483   36657  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36513   36483   36671  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36513   36483   36672  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36513   36483   36673  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36513   36483   36674  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36513   36483   36694  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36513   36483   36695  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S 65532      36513   36483   36696  0   13  80   0 - 318226 futex_ 07:13 ?       00:00:00 /usr/local/bin/kube-rbac-proxy --secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0
1 S root       36873       1   36873  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/218edc81822485123214796da8dcd3d0e53770cc343833080779a3eb3edbb8cb/userdata -c 218edc81822485123214796da8dcd3d0e53770cc343833080779a3eb3edbb8cb --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7fb4f565cd5xb48_aaf10555-293e-4c2b-baea-ffcdd4eeb046/manager/0.log --log-level info -n k8s_manager_openstack-baremetal-operator-controller-manager-7fb4f565cd5xb48_openstack-operators_aaf10555-293e-4c2b-baea-ffcdd4eeb046_0 -P /run/containers/storage/overlay-containers/218edc81822485123214796da8dcd3d0e53770cc343833080779a3eb3edbb8cb/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/218edc81822485123214796da8dcd3d0e53770cc343833080779a3eb3edbb8cb/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/218edc81822485123214796da8dcd3d0e53770cc343833080779a3eb3edbb8cb/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 218edc81822485123214796da8dcd3d0e53770cc343833080779a3eb3edbb8cb -s
4 S 1000660+   36875   36873   36875  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36883  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36884  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36885  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36886  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36887  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36888  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36889  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36890  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36891  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36892  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36893  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36894  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36895  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36896  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36897  0   17  80   0 - 615420 ep_pol 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 1000660+   36875   36873   36898  0   17  80   0 - 615420 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S root       37007       1   37007  0    1  80   0 -  2078 do_pol 07:13 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/a66aa4bf47787c389f1dbef338dbabaa03a848917e202d5fca4d6291465b6472/userdata -c a66aa4bf47787c389f1dbef338dbabaa03a848917e202d5fca4d6291465b6472 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack-operators_keystone-operator-controller-manager-6c9969c6c6-zvpws_21a68e29-471e-428d-9491-ae4b33a01e8a/manager/0.log --log-level info -n k8s_manager_keystone-operator-controller-manager-6c9969c6c6-zvpws_openstack-operators_21a68e29-471e-428d-9491-ae4b33a01e8a_0 -P /run/containers/storage/overlay-containers/a66aa4bf47787c389f1dbef338dbabaa03a848917e202d5fca4d6291465b6472/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/a66aa4bf47787c389f1dbef338dbabaa03a848917e202d5fca4d6291465b6472/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/a66aa4bf47787c389f1dbef338dbabaa03a848917e202d5fca4d6291465b6472/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u a66aa4bf47787c389f1dbef338dbabaa03a848917e202d5fca4d6291465b6472 -s
4 S 65532      37009   37007   37009  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37017  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37018  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:02 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37019  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37020  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37021  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      37009   37007   37022  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37023  0   17  80   0 - 632258 ep_pol 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37024  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37025  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37026  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
5 S 65532      37009   37007   37027  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37028  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37029  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37030  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37031  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:01 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S 65532      37009   37007   37032  0   17  80   0 - 632258 futex_ 07:13 ?       00:00:00 /manager --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect
1 S root       85034       1   85034  0    1  80   0 -  2078 do_pol 07:31 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/ee753f5b97138928ee8844a3bccd516454424ec5be6ac2dd907b904dc68f02e3/userdata -c ee753f5b97138928ee8844a3bccd516454424ec5be6ac2dd907b904dc68f02e3 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-marketplace_community-operators-k67nt_5e776c8b-d2f2-41b6-a795-dac7bf2c4a68/registry-server/0.log --log-level info -n k8s_registry-server_community-operators-k67nt_openshift-marketplace_5e776c8b-d2f2-41b6-a795-dac7bf2c4a68_0 -P /run/containers/storage/overlay-containers/ee753f5b97138928ee8844a3bccd516454424ec5be6ac2dd907b904dc68f02e3/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/ee753f5b97138928ee8844a3bccd516454424ec5be6ac2dd907b904dc68f02e3/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/ee753f5b97138928ee8844a3bccd516454424ec5be6ac2dd907b904dc68f02e3/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u ee753f5b97138928ee8844a3bccd516454424ec5be6ac2dd907b904dc68f02e3 -s
4 S 1000170+   85036   85034   85036  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85040  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
5 S 1000170+   85036   85034   85041  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85042  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85043  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85044  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85045  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85046  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85047  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85048  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85049  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85050  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85051  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85052  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85053  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85054  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85055  0   18  80   0 - 1439968 futex_ 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+   85036   85034   85060  0   18  80   0 - 1439968 ep_pol 07:31 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S root      110678       1  110678  0    1  80   0 -  2078 do_pol 07:57 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/def772fd14543962cbed67c0e5516a974f002f7dcd149b3fb0773829e3be4a72/userdata -c def772fd14543962cbed67c0e5516a974f002f7dcd149b3fb0773829e3be4a72 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-marketplace_certified-operators-4t2nc_6da6ea85-e420-40f6-829c-5983a746b478/registry-server/0.log --log-level info -n k8s_registry-server_certified-operators-4t2nc_openshift-marketplace_6da6ea85-e420-40f6-829c-5983a746b478_0 -P /run/containers/storage/overlay-containers/def772fd14543962cbed67c0e5516a974f002f7dcd149b3fb0773829e3be4a72/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/def772fd14543962cbed67c0e5516a974f002f7dcd149b3fb0773829e3be4a72/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/def772fd14543962cbed67c0e5516a974f002f7dcd149b3fb0773829e3be4a72/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u def772fd14543962cbed67c0e5516a974f002f7dcd149b3fb0773829e3be4a72 -s
4 S 1000170+  110680  110678  110680  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110684  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110685  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110686  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110687  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110688  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110689  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110690  0   18  80   0 - 1423567 ep_pol 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110691  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:00 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
5 S 1000170+  110680  110678  110692  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110693  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110694  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110695  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110696  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110697  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110698  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110699  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
1 S 1000170+  110680  110678  110758  0   18  80   0 - 1423567 futex_ 07:57 ?      00:00:01 /bin/opm serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache
0 S root      135087    5483  135087  0    1  80   0 -  1233 hrtime 08:19 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 3600
1 I root      152659       2  152659  0    1  80   0 -     0 worker 08:36 ?        00:00:00 [kworker/7:1-events]
1 I root      152811       2  152811  0    1  80   0 -     0 worker 08:36 ?        00:00:00 [kworker/4:1-events]
1 I root      160452       2  160452  0    1  80   0 -     0 worker 08:40 ?        00:00:00 [kworker/6:3-xfs-conv/vda4]
1 I root      162115       2  162115  0    1  80   0 -     0 worker 08:41 ?        00:00:00 [kworker/3:0-rcu_gp]
1 I root      177127       2  177127  0    1  80   0 -     0 worker 08:50 ?        00:00:00 [kworker/2:3-events]
1 S root      177565       1  177565  0    1  80   0 -  2078 do_pol 08:50 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/448f08b3dc3c0ff8ec372ea2140895a01da8a963041f286cc742e69bcb16dc40/userdata -c 448f08b3dc3c0ff8ec372ea2140895a01da8a963041f286cc742e69bcb16dc40 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack_ovsdbserver-nb-1_deb66def-1afa-46cc-bacf-2f6d3f8ed245/ovsdbserver-nb/0.log --log-level info -n k8s_ovsdbserver-nb_ovsdbserver-nb-1_openstack_deb66def-1afa-46cc-bacf-2f6d3f8ed245_0 -P /run/containers/storage/overlay-containers/448f08b3dc3c0ff8ec372ea2140895a01da8a963041f286cc742e69bcb16dc40/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/448f08b3dc3c0ff8ec372ea2140895a01da8a963041f286cc742e69bcb16dc40/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/448f08b3dc3c0ff8ec372ea2140895a01da8a963041f286cc742e69bcb16dc40/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 448f08b3dc3c0ff8ec372ea2140895a01da8a963041f286cc742e69bcb16dc40 -s
1 S root      177567       1  177567  0    1  80   0 -  2078 do_pol 08:50 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/48c4bc39465880c3fd3ff215e7973d374f3ac5c9112df09a363894883ff9fb64/userdata -c 48c4bc39465880c3fd3ff215e7973d374f3ac5c9112df09a363894883ff9fb64 --exit-dir /var/run/crio/exits -l /var/log/pods/openstack_ovsdbserver-nb-2_f8d17221-136d-42b3-9e68-f5eb5b026ae2/ovsdbserver-nb/0.log --log-level info -n k8s_ovsdbserver-nb_ovsdbserver-nb-2_openstack_f8d17221-136d-42b3-9e68-f5eb5b026ae2_0 -P /run/containers/storage/overlay-containers/48c4bc39465880c3fd3ff215e7973d374f3ac5c9112df09a363894883ff9fb64/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/48c4bc39465880c3fd3ff215e7973d374f3ac5c9112df09a363894883ff9fb64/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/48c4bc39465880c3fd3ff215e7973d374f3ac5c9112df09a363894883ff9fb64/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 48c4bc39465880c3fd3ff215e7973d374f3ac5c9112df09a363894883ff9fb64 -s
4 S 1000650+  177571  177565  177571  0    1  80   0 -   628 do_sig 08:50 ?        00:00:00 /usr/bin/dumb-init /usr/local/bin/container-scripts/setup.sh
4 S 1000650+  177572  177567  177572  0    1  80   0 -   628 do_sig 08:50 ?        00:00:00 /usr/bin/dumb-init /usr/local/bin/container-scripts/setup.sh
4 S 1000650+  177584  177572  177584  0    1  80   0 -  1811 do_wai 08:50 ?        00:00:00 bash /usr/local/bin/container-scripts/setup.sh
4 S 1000650+  177595  177571  177595  0    1  80   0 -  1811 do_wai 08:50 ?        00:00:00 bash /usr/local/bin/container-scripts/setup.sh
4 S 1000650+  177613  177584  177613  0    4  80   0 - 62221 do_pol 08:50 ?        00:00:00 ovsdb-server -vconsole:info --log-file=/dev/null --remote=punix:/tmp/ovnnb_db.sock --pidfile=/tmp/ovnnb_db.pid --unixctl=/tmp/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=/etc/pki/tls/private/ovndb.key --certificate=/etc/pki/tls/certs/ovndb.crt --ca-cert=/etc/pki/tls/certs/ovndbca.crt --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers -vfile:off /etc/ovn/ovnnb_db.db
1 S 1000650+  177613  177584  178084  0    4  80   0 - 62221 ep_pol 08:50 ?        00:00:00 ovsdb-server -vconsole:info --log-file=/dev/null --remote=punix:/tmp/ovnnb_db.sock --pidfile=/tmp/ovnnb_db.pid --unixctl=/tmp/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=/etc/pki/tls/private/ovndb.key --certificate=/etc/pki/tls/certs/ovndb.crt --ca-cert=/etc/pki/tls/certs/ovndbca.crt --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers -vfile:off /etc/ovn/ovnnb_db.db
1 S 1000650+  177613  177584  179553  0    4  80   0 - 62221 do_pol 08:50 ?        00:00:00 ovsdb-server -vconsole:info --log-file=/dev/null --remote=punix:/tmp/ovnnb_db.sock --pidfile=/tmp/ovnnb_db.pid --unixctl=/tmp/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=/etc/pki/tls/private/ovndb.key --certificate=/etc/pki/tls/certs/ovndb.crt --ca-cert=/etc/pki/tls/certs/ovndbca.crt --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers -vfile:off /etc/ovn/ovnnb_db.db
1 S 1000650+  177613  177584  179554  0    4  80   0 - 62221 do_pol 08:50 ?        00:00:00 ovsdb-server -vconsole:info --log-file=/dev/null --remote=punix:/tmp/ovnnb_db.sock --pidfile=/tmp/ovnnb_db.pid --unixctl=/tmp/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=/etc/pki/tls/private/ovndb.key --certificate=/etc/pki/tls/certs/ovndb.crt --ca-cert=/etc/pki/tls/certs/ovndbca.crt --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers -vfile:off /etc/ovn/ovnnb_db.db
4 S 1000650+  177633  177595  177633  0    4  80   0 - 62253 do_pol 08:50 ?        00:00:00 ovsdb-server -vconsole:info --log-file=/dev/null --remote=punix:/tmp/ovnnb_db.sock --pidfile=/tmp/ovnnb_db.pid --unixctl=/tmp/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=/etc/pki/tls/private/ovndb.key --certificate=/etc/pki/tls/certs/ovndb.crt --ca-cert=/etc/pki/tls/certs/ovndbca.crt --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers -vfile:off /etc/ovn/ovnnb_db.db
1 S 1000650+  177633  177595  178097  0    4  80   0 - 62253 ep_pol 08:50 ?        00:00:00 ovsdb-server -vconsole:info --log-file=/dev/null --remote=punix:/tmp/ovnnb_db.sock --pidfile=/tmp/ovnnb_db.pid --unixctl=/tmp/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=/etc/pki/tls/private/ovndb.key --certificate=/etc/pki/tls/certs/ovndb.crt --ca-cert=/etc/pki/tls/certs/ovndbca.crt --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers -vfile:off /etc/ovn/ovnnb_db.db
1 S 1000650+  177633  177595  179556  0    4  80   0 - 62253 do_pol 08:50 ?        00:00:00 ovsdb-server -vconsole:info --log-file=/dev/null --remote=punix:/tmp/ovnnb_db.sock --pidfile=/tmp/ovnnb_db.pid --unixctl=/tmp/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=/etc/pki/tls/private/ovndb.key --certificate=/etc/pki/tls/certs/ovndb.crt --ca-cert=/etc/pki/tls/certs/ovndbca.crt --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers -vfile:off /etc/ovn/ovnnb_db.db
1 S 1000650+  177633  177595  179557  0    4  80   0 - 62253 do_pol 08:50 ?        00:00:00 ovsdb-server -vconsole:info --log-file=/dev/null --remote=punix:/tmp/ovnnb_db.sock --pidfile=/tmp/ovnnb_db.pid --unixctl=/tmp/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=/etc/pki/tls/private/ovndb.key --certificate=/etc/pki/tls/certs/ovndb.crt --ca-cert=/etc/pki/tls/certs/ovndbca.crt --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers -vfile:off /etc/ovn/ovnnb_db.db
1 S root      178135       1  178135  0    1  80   0 -  2078 do_pol 08:50 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/6a2308089754012a0d926b6951c3bd0e3e5f81c20c41c3aafcb29ae66c3dcfbc/userdata -c 6a2308089754012a0d926b6951c3bd0e3e5f81c20c41c3aafcb29ae66c3dcfbc --exit-dir /var/run/crio/exits -l /var/log/pods/openstack_ovsdbserver-sb-2_d41676c8-318c-4033-a53e-c383f4f18fd7/ovsdbserver-sb/0.log --log-level info -n k8s_ovsdbserver-sb_ovsdbserver-sb-2_openstack_d41676c8-318c-4033-a53e-c383f4f18fd7_0 -P /run/containers/storage/overlay-containers/6a2308089754012a0d926b6951c3bd0e3e5f81c20c41c3aafcb29ae66c3dcfbc/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/6a2308089754012a0d926b6951c3bd0e3e5f81c20c41c3aafcb29ae66c3dcfbc/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/6a2308089754012a0d926b6951c3bd0e3e5f81c20c41c3aafcb29ae66c3dcfbc/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 6a2308089754012a0d926b6951c3bd0e3e5f81c20c41c3aafcb29ae66c3dcfbc -s
4 S 1000650+  178138  178135  178138  0    1  80   0 -   628 do_sig 08:50 ?        00:00:00 /usr/bin/dumb-init /usr/local/bin/container-scripts/setup.sh
4 S 1000650+  178159  178138  178159  0    1  80   0 -  1811 do_wai 08:50 ?        00:00:00 bash /usr/local/bin/container-scripts/setup.sh
4 S 1000650+  178182  178159  178182  0    4  80   0 - 62913 do_pol 08:50 ?        00:00:00 ovsdb-server -vconsole:info --log-file=/dev/null --remote=punix:/tmp/ovnsb_db.sock --pidfile=/tmp/ovnsb_db.pid --unixctl=/tmp/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=/etc/pki/tls/private/ovndb.key --certificate=/etc/pki/tls/certs/ovndb.crt --ca-cert=/etc/pki/tls/certs/ovndbca.crt --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers -vfile:off /etc/ovn/ovnsb_db.db
1 S 1000650+  178182  178159  178608  0    4  80   0 - 62913 ep_pol 08:50 ?        00:00:00 ovsdb-server -vconsole:info --log-file=/dev/null --remote=punix:/tmp/ovnsb_db.sock --pidfile=/tmp/ovnsb_db.pid --unixctl=/tmp/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=/etc/pki/tls/private/ovndb.key --certificate=/etc/pki/tls/certs/ovndb.crt --ca-cert=/etc/pki/tls/certs/ovndbca.crt --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers -vfile:off /etc/ovn/ovnsb_db.db
1 S 1000650+  178182  178159  179582  0    4  80   0 - 62913 do_pol 08:50 ?        00:00:00 ovsdb-server -vconsole:info --log-file=/dev/null --remote=punix:/tmp/ovnsb_db.sock --pidfile=/tmp/ovnsb_db.pid --unixctl=/tmp/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=/etc/pki/tls/private/ovndb.key --certificate=/etc/pki/tls/certs/ovndb.crt --ca-cert=/etc/pki/tls/certs/ovndbca.crt --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers -vfile:off /etc/ovn/ovnsb_db.db
1 S 1000650+  178182  178159  179583  0    4  80   0 - 62913 do_pol 08:50 ?        00:00:00 ovsdb-server -vconsole:info --log-file=/dev/null --remote=punix:/tmp/ovnsb_db.sock --pidfile=/tmp/ovnsb_db.pid --unixctl=/tmp/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=/etc/pki/tls/private/ovndb.key --certificate=/etc/pki/tls/certs/ovndb.crt --ca-cert=/etc/pki/tls/certs/ovndbca.crt --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers -vfile:off /etc/ovn/ovnsb_db.db
1 I root      182915       2  182915  0    1  80   0 -     0 worker 08:51 ?        00:00:00 [kworker/8:4-xfs-conv/vda4]
1 I root      185515       2  185515  0    1  80   0 -     0 worker 08:52 ?        00:00:00 [kworker/9:2-xfs-conv/vda4]
1 I root      193121       2  193121  0    1  80   0 -     0 worker 08:55 ?        00:00:00 [kworker/3:1-ipv6_addrconf]
1 I root      197218       2  197218  0    1  80   0 -     0 worker 08:56 ?        00:00:00 [kworker/5:6-mld]
1 I root      199702       2  199702  0    1  80   0 -     0 worker 08:57 ?        00:00:01 [kworker/0:4-ata_sff]
1 I root      199741       2  199741  0    1  80   0 -     0 worker 08:57 ?        00:00:00 [kworker/11:3-events]
1 I root      199767       2  199767  0    1  80   0 -     0 worker 08:57 ?        00:00:00 [kworker/11:5-events]
1 I root      201337       2  201337  0    1  80   0 -     0 worker 08:57 ?        00:00:00 [kworker/u24:33-flush-252:0]
0 S 1001      203545    9934  203545  0    1  80   0 -  1233 hrtime 08:58 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 3600
1 I root      206789       2  206789  0    1  80   0 -     0 worker 08:59 ?        00:00:00 [kworker/9:3-xfs-conv/vda4]
1 I root      209165       2  209165  0    1  80   0 -     0 worker 08:59 ?        00:00:00 [kworker/3:2-events]
1 I root      215179       2  215179  0    1  80   0 -     0 worker 09:01 ?        00:00:00 [kworker/8:5-xfs-conv/vda4]
1 I root      216464       2  216464  0    1  80   0 -     0 worker 09:02 ?        00:00:00 [kworker/8:14-xfs-conv/vda4]
1 I root      222406       2  222406  0    1  80   0 -     0 worker 09:03 ?        00:00:00 [kworker/7:2-mld]
1 I root      227179       2  227179  0    1  80   0 -     0 worker 09:04 ?        00:00:00 [kworker/4:6-events]
1 I root      227180       2  227180  0    1  80   0 -     0 worker 09:04 ?        00:00:00 [kworker/4:7-events]
1 I root      227269       2  227269  0    1  80   0 -     0 worker 09:04 ?        00:00:00 [kworker/1:10-events]
1 I root      227270       2  227270  0    1  80   0 -     0 worker 09:04 ?        00:00:00 [kworker/1:11-events]
1 I root      230106       2  230106  0    1  80   0 -     0 worker 09:05 ?        00:00:00 [kworker/5:0-ata_sff]
1 I root      230648       2  230648  0    1  80   0 -     0 worker 09:05 ?        00:00:00 [kworker/2:4-inet_frag_wq]
1 I root      231692       2  231692  0    1  80   0 -     0 worker 09:05 ?        00:00:00 [kworker/0:3-events]
1 I root      231693       2  231693  0    1  80   0 -     0 worker 09:05 ?        00:00:00 [kworker/0:5-mld]
1 I root      233366       2  233366  0    1  80   0 -     0 worker 09:06 ?        00:00:00 [kworker/u24:39-flush-252:0]
1 I root      233367       2  233367  0    1  80   0 -     0 worker 09:06 ?        00:00:00 [kworker/u24:40-flush-252:0]
1 I root      233371       2  233371  0    1  80   0 -     0 worker 09:06 ?        00:00:00 [kworker/u24:44-flush-252:0]
1 I root      233373       2  233373  0    1  80   0 -     0 worker 09:06 ?        00:00:00 [kworker/u24:46-events_unbound]
1 I root      233375       2  233375  0    1  80   0 -     0 worker 09:06 ?        00:00:00 [kworker/u24:48-xfs-blockgc/vda4]
1 I root      233378       2  233378  0    1  80   0 -     0 worker 09:06 ?        00:00:00 [kworker/u24:51-events_unbound]
1 I root      235450       2  235450  0    1  80   0 -     0 worker 09:06 ?        00:00:00 [kworker/6:0-kthrotld]
1 S root      238432       1  238432  0    1  80   0 -  2078 do_pol 09:07 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/8986dc133e78bed79b014504d2689f07e4fd3f6a8360c06366bc5c10812f1d43/userdata -c 8986dc133e78bed79b014504d2689f07e4fd3f6a8360c06366bc5c10812f1d43 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-68d8dd4df-rv8xk_10a017e0-9508-4ba6-8107-f7eda7c39515/prometheus-operator-admission-webhook/0.log --log-level info -n k8s_prometheus-operator-admission-webhook_obo-prometheus-operator-admission-webhook-68d8dd4df-rv8xk_openshift-operators_10a017e0-9508-4ba6-8107-f7eda7c39515_0 -P /run/containers/storage/overlay-containers/8986dc133e78bed79b014504d2689f07e4fd3f6a8360c06366bc5c10812f1d43/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/8986dc133e78bed79b014504d2689f07e4fd3f6a8360c06366bc5c10812f1d43/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/8986dc133e78bed79b014504d2689f07e4fd3f6a8360c06366bc5c10812f1d43/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 8986dc133e78bed79b014504d2689f07e4fd3f6a8360c06366bc5c10812f1d43 -s
4 S nfsnobo+  238434  238432  238434  0    9  80   0 - 317424 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238434  238432  238450  0    9  80   0 - 317424 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238434  238432  238451  0    9  80   0 - 317424 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238434  238432  238452  0    9  80   0 - 317424 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
5 S nfsnobo+  238434  238432  238453  0    9  80   0 - 317424 ep_pol 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238434  238432  238454  0    9  80   0 - 317424 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238434  238432  238455  0    9  80   0 - 317424 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238434  238432  238456  0    9  80   0 - 317424 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238434  238432  238475  0    9  80   0 - 317424 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S root      238457       1  238457  0    1  80   0 -  2078 do_pol 09:07 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/48a4147f89566cca3d81f4d3c21b4d458013345f288aeb034c12f7199e370f82/userdata -c 48a4147f89566cca3d81f4d3c21b4d458013345f288aeb034c12f7199e370f82 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-operators_perses-operator-54bc95c9fb-lsjsz_70cc7fd9-2b69-41fc-be41-569b6c564807/perses-operator/0.log --log-level info -n k8s_perses-operator_perses-operator-54bc95c9fb-lsjsz_openshift-operators_70cc7fd9-2b69-41fc-be41-569b6c564807_0 -P /run/containers/storage/overlay-containers/48a4147f89566cca3d81f4d3c21b4d458013345f288aeb034c12f7199e370f82/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/48a4147f89566cca3d81f4d3c21b4d458013345f288aeb034c12f7199e370f82/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/48a4147f89566cca3d81f4d3c21b4d458013345f288aeb034c12f7199e370f82/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 48a4147f89566cca3d81f4d3c21b4d458013345f288aeb034c12f7199e370f82 -s
1 S root      238458       1  238458  0    1  80   0 -  2078 do_pol 09:07 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/243e40c820c9628c66d3c3ea6b45d515b1a581c4c5f730b13d3d80db1de42e5a/userdata -c 243e40c820c9628c66d3c3ea6b45d515b1a581c4c5f730b13d3d80db1de42e5a --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-68d8dd4df-nq646_322bc19c-a989-4073-bd2d-166be2c156d9/prometheus-operator-admission-webhook/0.log --log-level info -n k8s_prometheus-operator-admission-webhook_obo-prometheus-operator-admission-webhook-68d8dd4df-nq646_openshift-operators_322bc19c-a989-4073-bd2d-166be2c156d9_0 -P /run/containers/storage/overlay-containers/243e40c820c9628c66d3c3ea6b45d515b1a581c4c5f730b13d3d80db1de42e5a/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/243e40c820c9628c66d3c3ea6b45d515b1a581c4c5f730b13d3d80db1de42e5a/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/243e40c820c9628c66d3c3ea6b45d515b1a581c4c5f730b13d3d80db1de42e5a/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 243e40c820c9628c66d3c3ea6b45d515b1a581c4c5f730b13d3d80db1de42e5a -s
1 S root      238461       1  238461  0    1  80   0 -  2078 do_pol 09:07 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/c5b4b54986966bc1dea066bd9e4ea6c9a33dc37d3d5fa25d6d4bc82c3c3e64b3/userdata -c c5b4b54986966bc1dea066bd9e4ea6c9a33dc37d3d5fa25d6d4bc82c3c3e64b3 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-kg8d2_8cea4c70-0056-479e-8759-eda2be569ee6/prometheus-operator/0.log --log-level info -n k8s_prometheus-operator_obo-prometheus-operator-7c8cf85677-kg8d2_openshift-operators_8cea4c70-0056-479e-8759-eda2be569ee6_0 -P /run/containers/storage/overlay-containers/c5b4b54986966bc1dea066bd9e4ea6c9a33dc37d3d5fa25d6d4bc82c3c3e64b3/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/c5b4b54986966bc1dea066bd9e4ea6c9a33dc37d3d5fa25d6d4bc82c3c3e64b3/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/c5b4b54986966bc1dea066bd9e4ea6c9a33dc37d3d5fa25d6d4bc82c3c3e64b3/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u c5b4b54986966bc1dea066bd9e4ea6c9a33dc37d3d5fa25d6d4bc82c3c3e64b3 -s
1 S root      238462       1  238462  0    1  80   0 -  2078 do_pol 09:07 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/45da6d09a6e39ceb40a82888c0eaaf93fedbcc2ab80291a907cae8ec374d0f1a/userdata -c 45da6d09a6e39ceb40a82888c0eaaf93fedbcc2ab80291a907cae8ec374d0f1a --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-operators_observability-operator-cc5f78dfc-krn6x_81f029f4-aa74-4b25-a6b5-b50a5933feec/operator/0.log --log-level info -n k8s_operator_observability-operator-cc5f78dfc-krn6x_openshift-operators_81f029f4-aa74-4b25-a6b5-b50a5933feec_0 -P /run/containers/storage/overlay-containers/45da6d09a6e39ceb40a82888c0eaaf93fedbcc2ab80291a907cae8ec374d0f1a/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/45da6d09a6e39ceb40a82888c0eaaf93fedbcc2ab80291a907cae8ec374d0f1a/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/45da6d09a6e39ceb40a82888c0eaaf93fedbcc2ab80291a907cae8ec374d0f1a/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 45da6d09a6e39ceb40a82888c0eaaf93fedbcc2ab80291a907cae8ec374d0f1a -s
4 S 1000350+  238465  238457  238465  0   13  80   0 - 540935 futex_ 09:07 ?       00:00:00 /bin/manager
1 S 1000350+  238465  238457  238503  0   13  80   0 - 540935 futex_ 09:07 ?       00:00:00 /bin/manager
1 S 1000350+  238465  238457  238504  0   13  80   0 - 540935 futex_ 09:07 ?       00:00:00 /bin/manager
1 S 1000350+  238465  238457  238505  0   13  80   0 - 540935 futex_ 09:07 ?       00:00:00 /bin/manager
1 S 1000350+  238465  238457  238506  0   13  80   0 - 540935 futex_ 09:07 ?       00:00:00 /bin/manager
1 S 1000350+  238465  238457  238507  0   13  80   0 - 540935 futex_ 09:07 ?       00:00:00 /bin/manager
1 S 1000350+  238465  238457  238518  0   13  80   0 - 540935 futex_ 09:07 ?       00:00:00 /bin/manager
1 S 1000350+  238465  238457  238519  0   13  80   0 - 540935 futex_ 09:07 ?       00:00:00 /bin/manager
1 S 1000350+  238465  238457  238520  0   13  80   0 - 540935 futex_ 09:07 ?       00:00:00 /bin/manager
1 S 1000350+  238465  238457  238521  0   13  80   0 - 540935 futex_ 09:07 ?       00:00:00 /bin/manager
1 S 1000350+  238465  238457  238529  0   13  80   0 - 540935 futex_ 09:07 ?       00:00:00 /bin/manager
1 S 1000350+  238465  238457  238530  0   13  80   0 - 540935 futex_ 09:07 ?       00:00:00 /bin/manager
1 S 1000350+  238465  238457  238531  0   13  80   0 - 540935 ep_pol 09:07 ?       00:00:00 /bin/manager
4 S nfsnobo+  238466  238458  238466  0    9  80   0 - 317488 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238466  238458  238481  0    9  80   0 - 317488 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238466  238458  238482  0    9  80   0 - 317488 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238466  238458  238483  0    9  80   0 - 317488 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238466  238458  238484  0    9  80   0 - 317488 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238466  238458  238524  0    9  80   0 - 317488 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
5 S nfsnobo+  238466  238458  238532  0    9  80   0 - 317488 ep_pol 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238466  238458  238537  0    9  80   0 - 317488 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
1 S nfsnobo+  238466  238458  238550  0    9  80   0 - 317488 futex_ 09:07 ?       00:00:00 /bin/admission-webhook --web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key
4 S 1000350+  238467  238462  238467  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238508  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238509  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
5 S 1000350+  238467  238462  238510  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238511  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238512  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238515  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238516  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238517  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238525  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238560  0   17  80   0 - 634852 ep_pol 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238563  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238564  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238565  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238566  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238634  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
1 S 1000350+  238467  238462  238644  0   17  80   0 - 634852 futex_ 09:07 ?       00:00:00 /manager --namespace=openshift-operators --images=alertmanager=registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404 --images=prometheus=registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00 --images=thanos=registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8 --images=perses=registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223 --images=ui-dashboards=registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610 --images=ui-distributed-tracing=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a --images=ui-distributed-tracing-pf5=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3 --images=ui-distributed-tracing-pf4=registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed --images=ui-logging=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf --images=ui-logging-pf4=registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552 --images=ui-troubleshooting-panel=registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad --images=ui-monitoring=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1 --images=ui-monitoring-pf5=registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e --images=korrel8r=registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7 --images=health-analyzer=registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd --openshift.enabled=true
4 S nfsnobo+  238468  238461  238468  0    9  80   0 - 322422 futex_ 09:07 ?       00:00:01 /bin/operator --prometheus-config-reloader=registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0 --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator
1 S nfsnobo+  238468  238461  238552  0    9  80   0 - 322422 futex_ 09:07 ?       00:00:00 /bin/operator --prometheus-config-reloader=registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0 --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator
1 S nfsnobo+  238468  238461  238553  0    9  80   0 - 322422 futex_ 09:07 ?       00:00:00 /bin/operator --prometheus-config-reloader=registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0 --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator
1 S nfsnobo+  238468  238461  238554  0    9  80   0 - 322422 futex_ 09:07 ?       00:00:00 /bin/operator --prometheus-config-reloader=registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0 --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator
1 S nfsnobo+  238468  238461  238558  0    9  80   0 - 322422 futex_ 09:07 ?       00:00:00 /bin/operator --prometheus-config-reloader=registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0 --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator
1 S nfsnobo+  238468  238461  238562  0    9  80   0 - 322422 futex_ 09:07 ?       00:00:00 /bin/operator --prometheus-config-reloader=registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0 --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator
1 S nfsnobo+  238468  238461  238568  0    9  80   0 - 322422 futex_ 09:07 ?       00:00:00 /bin/operator --prometheus-config-reloader=registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0 --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator
1 S nfsnobo+  238468  238461  238569  0    9  80   0 - 322422 ep_pol 09:07 ?       00:00:01 /bin/operator --prometheus-config-reloader=registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0 --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator
1 S nfsnobo+  238468  238461  238637  0    9  80   0 - 322422 futex_ 09:07 ?       00:00:00 /bin/operator --prometheus-config-reloader=registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0 --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator
1 I root      241447       2  241447  0    1  80   0 -     0 worker 09:08 ?        00:00:00 [kworker/5:1-events]
1 I root      241948       2  241948  0    1  80   0 -     0 worker 09:08 ?        00:00:00 [kworker/9:0-xfs-conv/vda4]
1 I root      244143       2  244143  0    1  80   0 -     0 worker 09:08 ?        00:00:00 [kworker/3:3-xfs-inodegc/vda4]
1 I root      245154       2  245154  0    1  80   0 -     0 worker 09:08 ?        00:00:00 [kworker/7:4-events]
1 I root      245227       2  245227  0    1  80   0 -     0 worker 09:08 ?        00:00:00 [kworker/10:11-rcu_gp]
1 I root      245228       2  245228  0    1  80   0 -     0 worker 09:08 ?        00:00:00 [kworker/10:12-mm_percpu_wq]
1 I root      245229       2  245229  0    1  80   0 -     0 worker 09:08 ?        00:00:00 [kworker/10:13-events]
1 I root      245230       2  245230  0    1  80   0 -     0 worker 09:08 ?        00:00:00 [kworker/10:14-cgroup_destroy]
4 S root      250934       1  250934  0    4  90  10 - 103633 ep_pol 09:10 ?       00:00:00 rpm-ostree start-daemon
5 S root      250934       1  250987  0    4  90  10 - 103633 do_pol 09:10 ?       00:00:00 rpm-ostree start-daemon
1 S root      250934       1  250988  0    4  90  10 - 103633 do_pol 09:10 ?       00:00:00 rpm-ostree start-daemon
1 S root      250934       1  250989  0    4  90  10 - 103633 do_pol 09:10 ?       00:00:00 rpm-ostree start-daemon
1 I root      254941       2  254941  0    1  80   0 -     0 worker 09:12 ?        00:00:00 [kworker/11:0-mld]
4 S root      254975    4764  254975  0    1  80   0 -  2078 do_pol 09:12 ?        00:00:00 /usr/bin/conmon -c 48c4bc39465880c3fd3ff215e7973d374f3ac5c9112df09a363894883ff9fb64 -n k8s_ovsdbserver-nb_ovsdbserver-nb-2_openstack_f8d17221-136d-42b3-9e68-f5eb5b026ae2_0 -r /usr/bin/crun -p /tmp/pidfile1471101885 -e -l /tmp/crio-log-48c4bc39465880c3fd3ff215e7973d374f3ac5c9112df09a363894883ff9fb64264705903 --socket-dir-path /var/run/crio --log-level info --sync --log-global-size-max 16777216 -s --exec-process-spec /tmp/exec-process-2177140216 --runtime-arg --root=/run/crun
4 S root      254981    4764  254981  0    1  80   0 -  2078 do_pol 09:12 ?        00:00:00 /usr/bin/conmon -c 448f08b3dc3c0ff8ec372ea2140895a01da8a963041f286cc742e69bcb16dc40 -n k8s_ovsdbserver-nb_ovsdbserver-nb-1_openstack_deb66def-1afa-46cc-bacf-2f6d3f8ed245_0 -r /usr/bin/crun -p /tmp/pidfile1159190181 -e -l /tmp/crio-log-448f08b3dc3c0ff8ec372ea2140895a01da8a963041f286cc742e69bcb16dc403713337067 --socket-dir-path /var/run/crio --log-level info --sync --log-global-size-max 16777216 -s --exec-process-spec /tmp/exec-process-2715975926 --runtime-arg --root=/run/crun
4 S 1000650+  254986  254975  254986  0    1  80   0 -  1811 do_wai 09:12 ?        00:00:00 bash /usr/local/bin/container-scripts/cleanup.sh
4 S 1000650+  254989  254981  254989  0    1  80   0 -  1811 do_wai 09:12 ?        00:00:00 bash /usr/local/bin/container-scripts/cleanup.sh
4 S root      255034    4764  255034  0    1  80   0 -  2078 do_pol 09:12 ?        00:00:00 /usr/bin/conmon -c 6a2308089754012a0d926b6951c3bd0e3e5f81c20c41c3aafcb29ae66c3dcfbc -n k8s_ovsdbserver-sb_ovsdbserver-sb-2_openstack_d41676c8-318c-4033-a53e-c383f4f18fd7_0 -r /usr/bin/crun -p /tmp/pidfile238573647 -e -l /tmp/crio-log-6a2308089754012a0d926b6951c3bd0e3e5f81c20c41c3aafcb29ae66c3dcfbc2652939465 --socket-dir-path /var/run/crio --log-level info --sync --log-global-size-max 16777216 -s --exec-process-spec /tmp/exec-process-61053431 --runtime-arg --root=/run/crun
4 S 1000650+  255074  255034  255074  0    1  80   0 -  1811 do_wai 09:12 ?        00:00:00 bash /usr/local/bin/container-scripts/cleanup.sh
1 I root      255076       2  255076  0    1  80   0 -     0 worker 09:12 ?        00:00:00 [kworker/6:1-xfs-conv/vda4]
1 I root      255223       2  255223  0    1  80   0 -     0 worker 09:12 ?        00:00:00 [kworker/1:0-events]
1 I root      255231       2  255231  0    1  80   0 -     0 worker 09:12 ?        00:00:00 [kworker/3:4-inode_switch_wbs]
1 I root      255470       2  255470  0    1  80   0 -     0 worker 09:12 ?        00:00:00 [kworker/4:0-cgroup_destroy]
1 I root      255471       2  255471  0    1  80   0 -     0 worker 09:12 ?        00:00:00 [kworker/4:2-rcu_gp]
1 I root      255579       2  255579  0    1  80   0 -     0 worker 09:12 ?        00:00:00 [kworker/11:1-events]
1 I root      255714       2  255714  0    1  80   0 -     0 worker 09:12 ?        00:00:00 [kworker/9:1-xfs-conv/vda4]
1 I root      256293       2  256293  0    1  80   0 -     0 worker 09:12 ?        00:00:00 [kworker/4:3-events]
1 I root      256412       2  256412  0    1  80   0 -     0 worker 09:12 ?        00:00:00 [kworker/11:2-events]
1 I root      257142       2  257142  0    1  80   0 -     0 worker 09:12 ?        00:00:00 [kworker/0:0]
5 S 1000560+  257392    8949  257392  0    4  80   0 - 63631 ep_pol 09:12 ?        00:00:00 /usr/sbin/haproxy -f /var/lib/haproxy/conf/haproxy.config -p /var/lib/haproxy/run/haproxy.pid -x /var/lib/haproxy/run/haproxy.sock -sf 975
5 S 1000560+  257392    8949  257394  0    4  80   0 - 63631 ep_pol 09:12 ?        00:00:00 /usr/sbin/haproxy -f /var/lib/haproxy/conf/haproxy.config -p /var/lib/haproxy/run/haproxy.pid -x /var/lib/haproxy/run/haproxy.sock -sf 975
5 S 1000560+  257392    8949  257395  0    4  80   0 - 63631 ep_pol 09:12 ?        00:00:00 /usr/sbin/haproxy -f /var/lib/haproxy/conf/haproxy.config -p /var/lib/haproxy/run/haproxy.pid -x /var/lib/haproxy/run/haproxy.sock -sf 975
5 S 1000560+  257392    8949  257396  0    4  80   0 - 63631 ep_pol 09:12 ?        00:00:00 /usr/sbin/haproxy -f /var/lib/haproxy/conf/haproxy.config -p /var/lib/haproxy/run/haproxy.pid -x /var/lib/haproxy/run/haproxy.sock -sf 975
0 S root      261958    5436  261958  0    1  80   0 -  1233 hrtime 09:13 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 60
1 S root      262160       1  262160  0    1  80   0 -  2078 do_pol 09:14 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/53e7e01fd393982e9d5cf0fad1086628e6c23cb26ce4df1c3e2c3d2ddf76bcd6/userdata -c 53e7e01fd393982e9d5cf0fad1086628e6c23cb26ce4df1c3e2c3d2ddf76bcd6 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-machine-config-operator_machine-config-daemon-z8f25_e12d3cfb-2ba7-4eb6-b6b4-bfc4cec93930/machine-config-daemon/27.log --log-level info -n k8s_machine-config-daemon_machine-config-daemon-z8f25_openshift-machine-config-operator_e12d3cfb-2ba7-4eb6-b6b4-bfc4cec93930_27 -P /run/containers/storage/overlay-containers/53e7e01fd393982e9d5cf0fad1086628e6c23cb26ce4df1c3e2c3d2ddf76bcd6/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/53e7e01fd393982e9d5cf0fad1086628e6c23cb26ce4df1c3e2c3d2ddf76bcd6/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/53e7e01fd393982e9d5cf0fad1086628e6c23cb26ce4df1c3e2c3d2ddf76bcd6/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 53e7e01fd393982e9d5cf0fad1086628e6c23cb26ce4df1c3e2c3d2ddf76bcd6 -s
4 S root      262164  262160  262164  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
1 S root      262164  262160  262193  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
1 S root      262164  262160  262194  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
1 S root      262164  262160  262195  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
1 S root      262164  262160  262196  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
1 S root      262164  262160  262197  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
5 S root      262164  262160  262198  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
5 S root      262164  262160  262199  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
1 S root      262164  262160  262200  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
1 S root      262164  262160  262201  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
1 S root      262164  262160  262202  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
5 S root      262164  262160  262203  0   16  80   0 - 602072 ep_pol 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
1 S root      262164  262160  262204  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
5 S root      262164  262160  262221  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
5 S root      262164  262160  262222  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
1 S root      262164  262160  262225  0   16  80   0 - 602072 futex_ 09:14 ?       00:00:00 /run/bin/machine-config-daemon start --payload-version=4.18.1 --root-mount=/
1 S root      262779       1  262779  0    1  80   0 -  2078 do_pol 09:14 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/c62d333e05580bf32830f78dcdcfa4ab1052446c8767c5251b160ca79a5658d6/userdata -c c62d333e05580bf32830f78dcdcfa4ab1052446c8767c5251b160ca79a5658d6 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-must-gather-hfgcf_must-gather-6kcxw_1515b15c-c878-43e5-8048-aa5c72ac62a6/gather/0.log --log-level info -n k8s_gather_must-gather-6kcxw_openshift-must-gather-hfgcf_1515b15c-c878-43e5-8048-aa5c72ac62a6_0 -P /run/containers/storage/overlay-containers/c62d333e05580bf32830f78dcdcfa4ab1052446c8767c5251b160ca79a5658d6/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/c62d333e05580bf32830f78dcdcfa4ab1052446c8767c5251b160ca79a5658d6/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/c62d333e05580bf32830f78dcdcfa4ab1052446c8767c5251b160ca79a5658d6/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u c62d333e05580bf32830f78dcdcfa4ab1052446c8767c5251b160ca79a5658d6 -s
4 S root      262781  262779  262781  0    1  80   0 -  1050 do_wai 09:14 ?        00:00:00 /bin/bash -c  echo "volume percentage checker started....." while true; do disk_usage=$(du -s "/must-gather" | awk '{print $1}') disk_space=$(df -P "/must-gather" | awk 'NR==2 {print $2}') usage_percentage=$(( (disk_usage * 100) / disk_space )) echo "volume usage percentage $usage_percentage" if [ "$usage_percentage" -gt "30" ]; then ?echo "Disk usage exceeds the volume percentage of 30 for mounted directory. Exiting..." ?# kill gathering process in gather container to prevent disk to use more. ?pkill --signal SIGKILL -f ADDITIONAL_NAMESPACES=kuttl,openshift-storage,openshift-marketplace,openshift-operators,sushy-emulator,tobiko OPENSTACK_DATABASES=ALL SOS_EDPM=all SOS_DECOMPRESS=0 gather ?exit 1 fi sleep 5 done & ADDITIONAL_NAMESPACES=kuttl,openshift-storage,openshift-marketplace,openshift-operators,sushy-emulator,tobiko OPENSTACK_DATABASES=ALL SOS_EDPM=all SOS_DECOMPRESS=0 gather; sync
5 S root      262785  262781  262785  0    1  80   0 -  1050 -      09:14 ?        00:00:00 /bin/bash -c  echo "volume percentage checker started....." while true; do disk_usage=$(du -s "/must-gather" | awk '{print $1}') disk_space=$(df -P "/must-gather" | awk 'NR==2 {print $2}') usage_percentage=$(( (disk_usage * 100) / disk_space )) echo "volume usage percentage $usage_percentage" if [ "$usage_percentage" -gt "30" ]; then ?echo "Disk usage exceeds the volume percentage of 30 for mounted directory. Exiting..." ?# kill gathering process in gather container to prevent disk to use more. ?pkill --signal SIGKILL -f ADDITIONAL_NAMESPACES=kuttl,openshift-storage,openshift-marketplace,openshift-operators,sushy-emulator,tobiko OPENSTACK_DATABASES=ALL SOS_EDPM=all SOS_DECOMPRESS=0 gather ?exit 1 fi sleep 5 done & ADDITIONAL_NAMESPACES=kuttl,openshift-storage,openshift-marketplace,openshift-operators,sushy-emulator,tobiko OPENSTACK_DATABASES=ALL SOS_EDPM=all SOS_DECOMPRESS=0 gather; sync
0 S root      262786  262781  262786  0    1  80   0 -  1050 do_wai 09:14 ?        00:00:00 /bin/bash /usr/bin/gather
4 S root      262790  262786  262790  0    1  80   0 -  1277 pipe_r 09:14 ?        00:00:00 /bin/bash /usr/bin/gather_run
4 S root      262791  262786  262791  0    1  80   0 -  1236 pipe_r 09:14 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=tee /usr/bin/tee /must-gather/run.log
1 S root      262813       1  262813  0    1  80   0 -  2078 do_pol 09:14 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/2020354af3cc1cb9278eb834869df5cd2bf2b6bec9d9c3b97bae8bb252e19bf0/userdata -c 2020354af3cc1cb9278eb834869df5cd2bf2b6bec9d9c3b97bae8bb252e19bf0 --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-must-gather-hfgcf_must-gather-6kcxw_1515b15c-c878-43e5-8048-aa5c72ac62a6/copy/0.log --log-level info -n k8s_copy_must-gather-6kcxw_openshift-must-gather-hfgcf_1515b15c-c878-43e5-8048-aa5c72ac62a6_0 -P /run/containers/storage/overlay-containers/2020354af3cc1cb9278eb834869df5cd2bf2b6bec9d9c3b97bae8bb252e19bf0/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/2020354af3cc1cb9278eb834869df5cd2bf2b6bec9d9c3b97bae8bb252e19bf0/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/2020354af3cc1cb9278eb834869df5cd2bf2b6bec9d9c3b97bae8bb252e19bf0/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 2020354af3cc1cb9278eb834869df5cd2bf2b6bec9d9c3b97bae8bb252e19bf0 -s
4 S root      262816  262813  262816  0    1  80   0 -  1050 do_wai 09:14 ?        00:00:00 /bin/bash -c trap : TERM INT; sleep infinity & wait
4 S root      262840  262816  262840  0    1  80   0 -  1236 do_sys 09:14 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep infinity
1 S root      262878  262790  262878  0    1  80   0 -  1083 do_wai 09:14 ?        00:00:00 /bin/bash /usr/bin/gather_run
0 S root      262879  262878  262879  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262883  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262884  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262885  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262886  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262887  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262894  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262895  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262896  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262898  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262899  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262905  0   14  80   0 - 1628864 ep_pol 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262922  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      262879  262878  262923  0   14  80   0 - 1628864 futex_ 09:14 ?      00:00:00 oc debug node/crc -- chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
1 S root      266666       1  266666  0    1  80   0 -  2078 do_pol 09:14 ?        00:00:00 /usr/bin/conmon -b /run/containers/storage/overlay-containers/296ed0750056f73d3fc06e4f5a26871e9848e7739cf2c173227fdf91ece092be/userdata -c 296ed0750056f73d3fc06e4f5a26871e9848e7739cf2c173227fdf91ece092be --exit-dir /var/run/crio/exits -l /var/log/pods/openshift-must-gather-hfgcf_crc-debug-pbp8n_ed61e53d-ad4d-47ec-ac35-88da1003d469/container-00/0.log --log-level info -n k8s_container-00_crc-debug-pbp8n_openshift-must-gather-hfgcf_ed61e53d-ad4d-47ec-ac35-88da1003d469_0 -P /run/containers/storage/overlay-containers/296ed0750056f73d3fc06e4f5a26871e9848e7739cf2c173227fdf91ece092be/userdata/conmon-pidfile -p /run/containers/storage/overlay-containers/296ed0750056f73d3fc06e4f5a26871e9848e7739cf2c173227fdf91ece092be/userdata/pidfile --persist-dir /var/lib/containers/storage/overlay-containers/296ed0750056f73d3fc06e4f5a26871e9848e7739cf2c173227fdf91ece092be/userdata -r /usr/bin/crun --runtime-arg --root=/run/crun --socket-dir-path /var/run/crio --syslog -u 296ed0750056f73d3fc06e4f5a26871e9848e7739cf2c173227fdf91ece092be -s
4 S root      266668  266666  266668  0    1  80   0 -  1047 do_wai 09:14 ?        00:00:00 bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ;           rm -rf "/var/tmp/sos-osp" &&           mkdir -p "/var/tmp/sos-osp" &&           sudo podman rm --force toolbox-osp;            sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools &&           toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir="/var/tmp/sos-osp" &&           if [[ "$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf "/var/tmp/sos-osp/podlogs.tar.xz" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi
4 S root      266718       1  266718  3    1  80   0 -  7840 ep_pol 09:14 ?        00:00:00 /usr/lib/systemd/systemd --user
5 S root      266728  266718  266728  0    1  80   0 - 47633 do_sig 09:14 ?        00:00:00 (sd-pam)
4 S root      267478  266668  267478  0    1  80   0 -  1080 do_wai 09:14 ?        00:00:00 /usr/bin/bash /usr/bin/toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=/var/tmp/sos-osp
0 S root      267705   24231  267705  0    1  80   0 -  1233 hrtime 09:14 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 30
0 S 1001      267875    5809  267875  0    1  80   0 -  1233 hrtime 09:14 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 60
1 I root      267882       2  267882  0    1  80   0 -     0 worker 09:14 ?        00:00:00 [kworker/8:0-xfs-conv/vda4]
5 S root      268054  266666  268054  0    1  80   0 -  2078 do_pol 09:14 ?        00:00:00 /usr/bin/conmon --api-version 1 -c e31f420c008f1b212a762fff4026ffb5dd933fb2c4bd4c9f722cb467266101da -u e31f420c008f1b212a762fff4026ffb5dd933fb2c4bd4c9f722cb467266101da -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/e31f420c008f1b212a762fff4026ffb5dd933fb2c4bd4c9f722cb467266101da/userdata -p /run/containers/storage/overlay-containers/e31f420c008f1b212a762fff4026ffb5dd933fb2c4bd4c9f722cb467266101da/userdata/pidfile -n toolbox-osp --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/e31f420c008f1b212a762fff4026ffb5dd933fb2c4bd4c9f722cb467266101da --full-attach -s -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/e31f420c008f1b212a762fff4026ffb5dd933fb2c4bd4c9f722cb467266101da/userdata/oci-log -t --conmon-pidfile /run/containers/storage/overlay-containers/e31f420c008f1b212a762fff4026ffb5dd933fb2c4bd4c9f722cb467266101da/userdata/conmon.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.skip_mount_home=true --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg e31f420c008f1b212a762fff4026ffb5dd933fb2c4bd4c9f722cb467266101da
4 S root      268056  268054  268056  0    1  80   0 -  1210 do_wai 09:14 ?        00:00:00 /usr/bin/bash
4 S root      268075  267478  268075  0    1  80   0 -  4243 do_pol 09:14 ?        00:00:00 sudo podman attach toolbox-osp
4 S root      268087  268075  268087  0   13  80   0 - 542119 futex_ 09:14 ?       00:00:00 podman attach toolbox-osp
1 S root      268087  268075  268088  0   13  80   0 - 542119 futex_ 09:14 ?       00:00:00 podman attach toolbox-osp
1 S root      268087  268075  268089  0   13  80   0 - 542119 futex_ 09:14 ?       00:00:00 podman attach toolbox-osp
1 S root      268087  268075  268090  0   13  80   0 - 542119 ep_pol 09:14 ?       00:00:00 podman attach toolbox-osp
1 S root      268087  268075  268091  0   13  80   0 - 542119 futex_ 09:14 ?       00:00:00 podman attach toolbox-osp
1 S root      268087  268075  268092  0   13  80   0 - 542119 futex_ 09:14 ?       00:00:00 podman attach toolbox-osp
1 S root      268087  268075  268093  0   13  80   0 - 542119 futex_ 09:14 ?       00:00:00 podman attach toolbox-osp
1 S root      268087  268075  268094  0   13  80   0 - 542119 futex_ 09:14 ?       00:00:00 podman attach toolbox-osp
1 S root      268087  268075  268095  0   13  80   0 - 542119 futex_ 09:14 ?       00:00:00 podman attach toolbox-osp
1 S root      268087  268075  268096  0   13  80   0 - 542119 futex_ 09:14 ?       00:00:00 podman attach toolbox-osp
1 S root      268087  268075  268097  0   13  80   0 - 542119 futex_ 09:14 ?       00:00:00 podman attach toolbox-osp
1 S root      268087  268075  268098  0   13  80   0 - 542119 futex_ 09:14 ?       00:00:00 podman attach toolbox-osp
1 S root      268087  268075  268099  0   13  80   0 - 542119 futex_ 09:14 ?       00:00:00 podman attach toolbox-osp
4 S root      268100  268056  268100 10    9  80   0 - 162773 futex_ 09:14 ?       00:00:01 /usr/bin/python3 -s /usr/sbin/sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=/var/tmp/sos-osp
1 S root      268100  268056  270474  0    9  80   0 - 162773 futex_ 09:14 ?       00:00:00 /usr/bin/python3 -s /usr/sbin/sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=/var/tmp/sos-osp
1 S root      268100  268056  270475  0    9  80   0 - 162773 futex_ 09:14 ?       00:00:00 /usr/bin/python3 -s /usr/sbin/sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=/var/tmp/sos-osp
1 S root      268100  268056  270476  0    9  80   0 - 162773 futex_ 09:14 ?       00:00:00 /usr/bin/python3 -s /usr/sbin/sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=/var/tmp/sos-osp
1 S root      268100  268056  270477  0    9  80   0 - 162773 futex_ 09:14 ?       00:00:00 /usr/bin/python3 -s /usr/sbin/sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=/var/tmp/sos-osp
5 S root      268100  268056  270478  5    9  80   0 - 162773 pipe_r 09:14 ?       00:00:00 /usr/bin/python3 -s /usr/sbin/sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=/var/tmp/sos-osp
5 S root      268100  268056  270684 12    9  80   0 - 162773 do_sel 09:14 ?       00:00:00 /usr/bin/python3 -s /usr/sbin/sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=/var/tmp/sos-osp
1 S root      268100  268056  271116  2    9  80   0 - 162773 pipe_r 09:14 ?       00:00:00 /usr/bin/python3 -s /usr/sbin/sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=/var/tmp/sos-osp
5 S root      268100  268056  271266 50    9  80   0 - 162773 do_sel 09:14 ?       00:00:02 /usr/bin/python3 -s /usr/sbin/sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=/var/tmp/sos-osp
4 S root      269595    1300  269595  0    1  80   0 -  4493 wait_w 09:14 ?        00:00:00 sshd: [accepted]
1 I root      270272       2  270272  0    1  80   0 -     0 worker 09:14 ?        00:00:00 [kworker/6:2-xfs-conv/vda4]
1 I root      270273       2  270273  0    1  80   0 -     0 worker 09:14 ?        00:00:00 [kworker/6:4-xfs-conv/vda4]
1 I root      270279       2  270279  0    1  80   0 -     0 worker 09:14 ?        00:00:00 [kworker/6:5-events]
4 S root      270998  268100  270998  0    1  80   0 -   767 sigsus 09:14 ?        00:00:00 timeout 300s journalctl --no-pager --boot
0 R root      270999  270998  270999 87    1  80   0 - 137661 -     09:14 ?        00:00:04 journalctl --no-pager --boot
1 I root      272008       2  272008  0    1  80   0 -     0 worker 09:14 ?        00:00:00 [kworker/8:1-xfs-conv/vda4]
1 I root      272020       2  272020  0    1  80   0 -     0 worker 09:14 ?        00:00:00 [kworker/8:2-xfs-conv/vda4]
1 I root      272036       2  272036  0    1  80   0 -     0 worker 09:14 ?        00:00:00 [kworker/8:3-xfs-conv/vda4]
1 I root      272095       2  272095  0    1  80   0 -     0 worker 09:14 ?        00:00:00 [kworker/8:6-events]
1 I root      272101       2  272101  0    1  80   0 -     0 worker 09:14 ?        00:00:00 [kworker/9:4-xfs-conv/vda4]
1 I root      272102       2  272102  0    1  80   0 -     0 worker 09:14 ?        00:00:00 [kworker/8:7-events]
1 D root      272120       2  272120  0    1  80   0 -     0 rq_qos 09:14 ?        00:00:00 [kworker/9:5+kthrotld]
0 S 1000650+  273145  254986  273145  0    1  80   0 -  1964 hrtime 09:14 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 1
0 S 1000650+  273154  254989  273154  0    1  80   0 -  1964 hrtime 09:14 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 1
0 S 1000650+  273155  255074  273155  0    1  80   0 -  1964 hrtime 09:14 ?        00:00:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 1
4 S root      273352  268100  273352  0    1  80   0 -   767 sigsus 09:14 ?        00:00:00 timeout 300s ps -elfL
4 R root      273361  273352  273361  0    1  80   0 -  4063 -      09:14 ?        00:00:00 ps -elfL
